版本:Flink1.13.3、Hudi0.10.1、Hive3.1.2、Hadoop3.2.1
编译:Hudi:mvn clean package -DskipITs -Dmaven.test.skip=true
-Dhadoop.version=3.2.1 -Pflink-bundle-shade-hive3
Flink-SQL-HIVE:mvn clean install -Dfast -Dhadoop.version=3.2.1 -Dscala-2.11
-DskipTests -Dfast -T 4 -Dmaven.compile.fork=true -Dmaven.javadoc.skip=true
-Dcheckstyle.skip=true
启动:./sql-client.sh embedded -j ../lib/hudi-flink-bundle_2.11-0.10.1-rc1.jar
Lib目录:
[root@rhy-t-bd-java lib]# ll
total 271504
-rw-r--r-- 1 root root 92313 Oct 12 2021 flink-csv-1.13.3.jar
-rw-r--r-- 1 root root 106535831 Oct 12 2021 flink-dist_2.12-1.13.3.jar
-rw-r--r-- 1 root root 148127 Oct 12 2021 flink-json-1.13.3.jar
-rwxrwxrwx 1 root root 7709740 Jun 8 2021
flink-shaded-zookeeper-3.4.14.jar
-rw-r--r-- 1 stack wheel 48845196 Sep 13 18:43
flink-sql-connector-hive-3.1.2_2.11-1.13.2.jar
-rw-r--r-- 1 root root 35051553 Oct 12 2021 flink-table_2.12-1.13.3.jar
-rw-r--r-- 1 root root 38613339 Oct 12 2021
flink-table-blink_2.12-1.13.3.jar
-rw-r--r-- 1 root root 38955252 Sep 13 17:20
hudi-flink-bundle_2.11-0.10.1-rc1.jar
-rwxrwxrwx 1 root root 67114 Mar 31 2021 log4j-1.2-api-2.12.1.jar
-rwxrwxrwx 1 root root 276771 Mar 31 2021 log4j-api-2.12.1.jar
-rwxrwxrwx 1 root root 1674433 Mar 31 2021 log4j-core-2.12.1.jar
-rwxrwxrwx 1 root root 23518 Mar 31 2021 log4j-slf4j-impl-2.12.1.jar
Flink-SQL:CREATE TABLE paat_hudi_flink_test(
id bigint ,
name string,
birthday TIMESTAMP(3),
tsTIMESTAMP(3),
partition VARCHAR(20),
primary key(id) not enforced -- the uuid primary key must be specified
)
PARTITIONED BY (partition)
with(
'connector'='hudi',
'path' = 'hdfs://emr-cluster/user/hive/hudi/warehouse/ods_hudi.hudi_flink_test/'
, 'hoodie.datasource.write.recordkey.field' = 'id'
, 'write.precombine.field' = 'ts'
, 'write.tasks' = '1'
, 'compaction.tasks' = '1'
, 'write.rate.limit' = '2000'
, 'table.type' = 'MERGE_ON_READ'
, 'compaction.async.enable' = 'true'
, 'compaction.trigger.strategy' = 'num_commits'
, 'compaction.max_memory' = '1024'
, 'changelog.enable' = 'true'
, 'read.streaming.enable' = 'true'
, 'read.streaming.check-interval' = '4'
, 'hive_sync.enable' = 'true'
, 'hive_sync.mode'= 'hms'
, 'hive_sync.metastore.uris' = 'thrift://:9083'
, 'hive_sync.jdbc_url' = 'jdbc:hive2://
, 'hive_sync.jdbc_url' = 'jdbc:hive2://:10000'
, 'hive_sync.table' = 'hudi_flink_test'
, 'hive_sync.db' = 'ods_hudi'
, 'hive_sync.username' = '
, 'hive_sync.table' = 'hudi_flink_test'
, 'hive_sync.db' = 'ods_hudi'
, 'hive_sync.username' = ''
, 'hive_sync.password' = '
, 'hive_sync.password' = '*^'
, 'hive_sync.support_timestamp' = 'true'
);
Query: select * from paat_hudi_flink_test;
出现错误:2022-09-13 18:45:29,203 INFO
org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor [] - The
RpcEndpoint jobmanager_2 failed.
org.apache.flink.runtime.rpc.akka.exceptions.AkkaRpcException: Could not start
RpcEndpoint jobmanager_2.
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:610)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleControlMessage(AkkaRpcActor.java:180)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:26)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:21)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at scala.PartialFunction.applyOrElse(PartialFunction.scala:123)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:21)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.actor.Actor.aroundReceive(Actor.scala:517)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.actor.Actor.aroundReceive$(Actor.scala:515)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:225)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.actor.ActorCell.receiveMessage(ActorCell.scala:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.actor.ActorCell.invoke(ActorCell.scala:561)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:258)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.dispatch.Mailbox.run(Mailbox.scala:225)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.dispatch.Mailbox.exec(Mailbox.scala:235)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at
akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at
akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
[flink-dist_2.12-1.13.3.jar:1.13.3]
at
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
[flink-dist_2.12-1.13.3.jar:1.13.3]
Caused by: org.apache.flink.runtime.jobmaster.JobMasterException: Could not
start the JobMaster.
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:385)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
Caused by: org.apache.flink.util.FlinkRuntimeException: Failed to start the
operator coordinators
at
org.apache.flink.runtime.scheduler.DefaultOperatorCoordinatorHandler.startAllOperatorCoordinators(DefaultOperatorCoordinatorHandler.java:90)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.SchedulerBase.startScheduling(SchedulerBase.java:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startScheduling(JobMaster.java:955)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startJobExecution(JobMaster.java:873)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:383)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
Caused by: org.apache.hudi.exception.HoodieIOException: Failed to get instance
of org.apache.hadoop.fs.FileSystem
at org.apache.hudi.common.fs.FSUtils.getFs(FSUtils.java:104)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at org.apache.hudi.util.StreamerUtil.tableExists(StreamerUtil.java:288)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.util.StreamerUtil.initTableIfNotExists(StreamerUtil.java:258)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.sink.StreamWriteOperatorCoordinator.start(StreamWriteOperatorCoordinator.java:164)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.start(OperatorCoordinatorHolder.java:194)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.DefaultOperatorCoordinatorHandler.startAllOperatorCoordinators(DefaultOperatorCoordinatorHandler.java:85)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.SchedulerBase.startScheduling(SchedulerBase.java:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startScheduling(JobMaster.java:955)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startJobExecution(JobMaster.java:873)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:383)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
Caused by: java.io.IOException: Couldn't create proxy provider class
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:261)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:224)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithClientProtocol(NameNodeProxiesClient.java:134)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:356)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:290)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:171)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3303)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
~[hadoop-common-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3352)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3320)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:479)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:365)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hudi.common.fs.FSUtils.getFs(FSUtils.java:102)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at org.apache.hudi.util.StreamerUtil.tableExists(StreamerUtil.java:288)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.util.StreamerUtil.initTableIfNotExists(StreamerUtil.java:258)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.sink.StreamWriteOperatorCoordinator.start(StreamWriteOperatorCoordinator.java:164)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.start(OperatorCoordinatorHolder.java:194)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.DefaultOperatorCoordinatorHandler.startAllOperatorCoordinators(DefaultOperatorCoordinatorHandler.java:85)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.SchedulerBase.startScheduling(SchedulerBase.java:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startScheduling(JobMaster.java:955)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startJobExecution(JobMaster.java:873)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:383)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
Caused by: java.lang.reflect.InvocationTargetException
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[?:1.8.0_231]
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[?:1.8.0_231]
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[?:1.8.0_231]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
~[?:1.8.0_231]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:245)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:224)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithClientProtocol(NameNodeProxiesClient.java:134)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:356)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:290)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:171)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3303)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
~[hadoop-common-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3352)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3320)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:479)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:365)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hudi.common.fs.FSUtils.getFs(FSUtils.java:102)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at org.apache.hudi.util.StreamerUtil.tableExists(StreamerUtil.java:288)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.util.StreamerUtil.initTableIfNotExists(StreamerUtil.java:258)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.sink.StreamWriteOperatorCoordinator.start(StreamWriteOperatorCoordinator.java:164)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.start(OperatorCoordinatorHolder.java:194)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.DefaultOperatorCoordinatorHandler.startAllOperatorCoordinators(DefaultOperatorCoordinatorHandler.java:85)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.SchedulerBase.startScheduling(SchedulerBase.java:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startScheduling(JobMaster.java:955)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startJobExecution(JobMaster.java:873)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:383)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
Caused by: java.lang.NoSuchMethodError:
com.google.common.base.Preconditions.checkArgument(ZLjava/lang/String;Ljava/lang/Object;)V
at org.apache.hadoop.conf.Configuration.set(Configuration.java:1357)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.conf.Configuration.set(Configuration.java:1338)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1515)
~[hadoop-common-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.server.namenode.ha.AbstractNNFailoverProxyProvider.(AbstractNNFailoverProxyProvider.java:70)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.(ConfiguredFailoverProxyProvider.java:50)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider.(ConfiguredFailoverProxyProvider.java:45)
~[hadoop-hdfs-client-3.2.1.jar:?]
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[?:1.8.0_231]
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[?:1.8.0_231]
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[?:1.8.0_231]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
~[?:1.8.0_231]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:245)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createFailoverProxyProvider(NameNodeProxiesClient.java:224)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithClientProtocol(NameNodeProxiesClient.java:134)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:356)
~[hadoop-hdfs-client-3.2.1.jar:?]
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:290)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:171)
~[hadoop-hdfs-client-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3303)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124)
~[hadoop-common-3.2.1.jar:?]
at
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3352)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3320)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:479)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hadoop.fs.Path.getFileSystem(Path.java:365)
~[hadoop-common-3.2.1.jar:?]
at org.apache.hudi.common.fs.FSUtils.getFs(FSUtils.java:102)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at org.apache.hudi.util.StreamerUtil.tableExists(StreamerUtil.java:288)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.util.StreamerUtil.initTableIfNotExists(StreamerUtil.java:258)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.hudi.sink.StreamWriteOperatorCoordinator.start(StreamWriteOperatorCoordinator.java:164)
~[hudi-flink-bundle_2.11-0.10.1-rc1.jar:0.10.1-rc1]
at
org.apache.flink.runtime.operators.coordination.OperatorCoordinatorHolder.start(OperatorCoordinatorHolder.java:194)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.DefaultOperatorCoordinatorHandler.startAllOperatorCoordinators(DefaultOperatorCoordinatorHandler.java:85)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.scheduler.SchedulerBase.startScheduling(SchedulerBase.java:592)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startScheduling(JobMaster.java:955)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.startJobExecution(JobMaster.java:873)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.jobmaster.JobMaster.onStart(JobMaster.java:383)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.RpcEndpoint.internalCallOnStart(RpcEndpoint.java:181)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
at
org.apache.flink.runtime.rpc.akka.AkkaRpcActor$StoppedState.start(AkkaRpcActor.java:605)
~[flink-dist_2.12-1.13.3.jar:1.13.3]
... 20 more
请问这是guava版本不一致造成的吗??有什么解决方案吗??或者是其它什么原因??