masonone opened a new issue #2015:
URL: https://github.com/apache/iceberg/issues/2015


   ```
   Fail to run sql command: SELECT * FROM sample
   org.apache.flink.table.api.TableException: Failed to execute sql
        at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:749)
        at 
org.apache.flink.table.api.internal.TableImpl.execute(TableImpl.java:570)
        at 
org.apache.zeppelin.flink.Flink111Shims.collectToList(Flink111Shims.java:174)
        at 
org.apache.zeppelin.flink.FlinkZeppelinContext.showData(FlinkZeppelinContext.scala:115)
        at 
org.apache.zeppelin.interpreter.ZeppelinContext.showData(ZeppelinContext.java:67)
        at 
org.apache.zeppelin.flink.FlinkBatchSqlInterpreter.callInnerSelect(FlinkBatchSqlInterpreter.java:60)
        at 
org.apache.zeppelin.flink.FlinkSqlInterrpeter.callSelect(FlinkSqlInterrpeter.java:494)
        at 
org.apache.zeppelin.flink.FlinkSqlInterrpeter.callCommand(FlinkSqlInterrpeter.java:265)
        at 
org.apache.zeppelin.flink.FlinkSqlInterrpeter.runSqlList(FlinkSqlInterrpeter.java:159)
        at 
org.apache.zeppelin.flink.FlinkSqlInterrpeter.interpret(FlinkSqlInterrpeter.java:124)
        at 
org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:110)
        at 
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:776)
        at 
org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
        at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
        at 
org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
        at 
org.apache.zeppelin.scheduler.ParallelScheduler.lambda$runJobInScheduler$0(ParallelScheduler.java:39)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)
   Caused by: org.apache.flink.util.FlinkException: Failed to execute job 
'collect'.
        at 
org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1829)
        at 
org.apache.flink.api.java.ScalaShellStreamEnvironment.executeAsync(ScalaShellStreamEnvironment.java:75)
        at 
org.apache.flink.table.planner.delegation.ExecutorBase.executeAsync(ExecutorBase.java:57)
        at 
org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:738)
        ... 18 more
   Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed to 
submit JobGraph.
        at 
org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$7(RestClusterClient.java:366)
        at 
java.util.concurrent.CompletableFuture.uniExceptionally(CompletableFuture.java:884)
        at 
java.util.concurrent.CompletableFuture$UniExceptionally.tryFire(CompletableFuture.java:866)
        at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at 
java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:1990)
        at 
org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$8(FutureUtils.java:292)
        at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
        at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
        at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488)
        at 
java.util.concurrent.CompletableFuture.postFire(CompletableFuture.java:575)
        at 
java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:943)
        at 
java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
        ... 3 more
   Caused by: org.apache.flink.runtime.rest.util.RestClientException: [Internal 
server error., <Exception on server side:
   org.apache.flink.runtime.client.JobSubmissionException: Failed to submit job.
        at 
org.apache.flink.runtime.dispatcher.Dispatcher.lambda$internalSubmitJob$3(Dispatcher.java:362)
        at 
java.util.concurrent.CompletableFuture.uniHandle(CompletableFuture.java:836)
        at 
java.util.concurrent.CompletableFuture$UniHandle.tryFire(CompletableFuture.java:811)
        at 
java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:456)
        at akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40)
        at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44)
        at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at 
akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at 
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)
   Caused by: org.apache.flink.runtime.client.JobExecutionException: Could not 
instantiate JobManager.
        at 
org.apache.flink.runtime.dispatcher.Dispatcher.lambda$createJobManagerRunner$6(Dispatcher.java:427)
        at 
java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1604)
        ... 6 more
   Caused by: org.apache.flink.runtime.JobException: Creating the input splits 
caused an error: Stack map does not match the one at exception handler 69
   Exception Details:
     Location:
       
org/apache/iceberg/hive/HiveCatalog.loadNamespaceMetadata(Lorg/apache/iceberg/catalog/Namespace;)Ljava/util/Map;
 @69: astore_2
     Reason:
       Type 'org/apache/hadoop/hive/metastore/api/NoSuchObjectException' 
(current frame, stack[0]) is not assignable to 'org/apache/thrift/TException' 
(stack map, stack[0])
     Current Frame:
       bci: @26
       flags: { }
       locals: { 'org/apache/iceberg/hive/HiveCatalog', 
'org/apache/iceberg/catalog/Namespace' }
       stack: { 'org/apache/hadoop/hive/metastore/api/NoSuchObjectException' }
     Stackmap Frame:
       bci: @69
       flags: { }
       locals: { 'org/apache/iceberg/hive/HiveCatalog', 
'org/apache/iceberg/catalog/Namespace' }
       stack: { 'org/apache/thrift/TException' }
     Bytecode:
       0x0000000: 2a2b b700 759a 0015 bb00 c759 12c9 04bd
       0x0000010: 00cb 5903 2b53 b700 cebf 2ab4 0038 2bba
       0x0000020: 0236 0000 b600 9ac0 0238 4d2a 2cb7 023c
       0x0000030: 4eb2 00bd 1302 3e2b 2db9 0204 0100 b900
       0x0000040: c504 002d b04d bb00 c759 2c12 c904 bd00
       0x0000050: cb59 032b 53b7 0229 bf4d bb00 d059 bb00
       0x0000060: d259 b700 d313 022b b600 d92b b600 dc13
       0x0000070: 01a4 b600 d9b6 00e0 2cb7 00e3 bf4d b800
       0x0000080: 40b6 00e6 bb00 d059 bb00 d259 b700 d313
       0x0000090: 022d b600 d92b b600 dc13 01a4 b600 d9b6
       0x00000a0: 00e0 2cb7 00e3 bf                      
     Exception Handler Table:
       bci [26, 68] => handler: 69
       bci [26, 68] => handler: 69
       bci [26, 68] => handler: 89
       bci [26, 68] => handler: 125
     Stackmap Table:
       same_frame(@26)
       same_locals_1_stack_item_frame(@69,Object[#111])
       same_locals_1_stack_item_frame(@89,Object[#111])
       same_locals_1_stack_item_frame(@125,Object[#113])
   
        at 
org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:272)
        at 
org.apache.flink.runtime.executiongraph.ExecutionGraph.attachJobGraph(ExecutionGraph.java:814)
        at 
org.apache.flink.runtime.executiongraph.ExecutionGraphBuilder.buildGraph(ExecutionGraphBuilder.java:228)
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.createExecutionGraph(SchedulerBase.java:270)
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.createAndRestoreExecutionGraph(SchedulerBase.java:244)
        at 
org.apache.flink.runtime.scheduler.SchedulerBase.<init>(SchedulerBase.java:231)
        at 
org.apache.flink.runtime.scheduler.DefaultScheduler.<init>(DefaultScheduler.java:119)
        at 
org.apache.flink.runtime.scheduler.DefaultSchedulerFactory.createInstance(DefaultSchedulerFactory.java:103)
        at 
org.apache.flink.runtime.jobmaster.JobMaster.createScheduler(JobMaster.java:290)
        at 
org.apache.flink.runtime.jobmaster.JobMaster.<init>(JobMaster.java:278)
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.createJobMasterService(DefaultJobMasterServiceFactory.java:98)
        at 
org.apache.flink.runtime.jobmaster.factories.DefaultJobMasterServiceFactory.createJobMasterService(DefaultJobMasterServiceFactory.java:40)
        at 
org.apache.flink.runtime.jobmaster.JobManagerRunnerImpl.<init>(JobManagerRunnerImpl.java:140)
        at 
org.apache.flink.runtime.dispatcher.DefaultJobManagerRunnerFactory.createJobManagerRunner(DefaultJobManagerRunnerFactory.java:84)
        at 
org.apache.flink.runtime.dispatcher.Dispatcher.lambda$createJobManagerRunner$6(Dispatcher.java:417)
        ... 7 more
   Caused by: java.lang.VerifyError: Stack map does not match the one at 
exception handler 69
   Exception Details:
     Location:
       
org/apache/iceberg/hive/HiveCatalog.loadNamespaceMetadata(Lorg/apache/iceberg/catalog/Namespace;)Ljava/util/Map;
 @69: astore_2
     Reason:
       Type 'org/apache/hadoop/hive/metastore/api/NoSuchObjectException' 
(current frame, stack[0]) is not assignable to 'org/apache/thrift/TException' 
(stack map, stack[0])
     Current Frame:
       bci: @26
       flags: { }
       locals: { 'org/apache/iceberg/hive/HiveCatalog', 
'org/apache/iceberg/catalog/Namespace' }
       stack: { 'org/apache/hadoop/hive/metastore/api/NoSuchObjectException' }
     Stackmap Frame:
       bci: @69
       flags: { }
       locals: { 'org/apache/iceberg/hive/HiveCatalog', 
'org/apache/iceberg/catalog/Namespace' }
       stack: { 'org/apache/thrift/TException' }
     Bytecode:
       0x0000000: 2a2b b700 759a 0015 bb00 c759 12c9 04bd
       0x0000010: 00cb 5903 2b53 b700 cebf 2ab4 0038 2bba
       0x0000020: 0236 0000 b600 9ac0 0238 4d2a 2cb7 023c
       0x0000030: 4eb2 00bd 1302 3e2b 2db9 0204 0100 b900
       0x0000040: c504 002d b04d bb00 c759 2c12 c904 bd00
       0x0000050: cb59 032b 53b7 0229 bf4d bb00 d059 bb00
       0x0000060: d259 b700 d313 022b b600 d92b b600 dc13
       0x0000070: 01a4 b600 d9b6 00e0 2cb7 00e3 bf4d b800
       0x0000080: 40b6 00e6 bb00 d059 bb00 d259 b700 d313
       0x0000090: 022d b600 d92b b600 dc13 01a4 b600 d9b6
       0x00000a0: 00e0 2cb7 00e3 bf                      
     Exception Handler Table:
       bci [26, 68] => handler: 69
       bci [26, 68] => handler: 69
       bci [26, 68] => handler: 89
       bci [26, 68] => handler: 125
     Stackmap Table:
       same_frame(@26)
       same_locals_1_stack_item_frame(@69,Object[#111])
       same_locals_1_stack_item_frame(@89,Object[#111])
       same_locals_1_stack_item_frame(@125,Object[#113])
   
        at 
org.apache.iceberg.flink.CatalogLoader$HiveCatalogLoader.loadCatalog(CatalogLoader.java:95)
        at 
org.apache.iceberg.flink.TableLoader$CatalogTableLoader.open(TableLoader.java:108)
        at 
org.apache.iceberg.flink.source.FlinkInputFormat.createInputSplits(FlinkInputFormat.java:75)
        at 
org.apache.iceberg.flink.source.FlinkInputFormat.createInputSplits(FlinkInputFormat.java:40)
        at 
org.apache.flink.runtime.executiongraph.ExecutionJobVertex.<init>(ExecutionJobVertex.java:258)
        ... 21 more
   
   End of exception on server side>]
        at 
org.apache.flink.runtime.rest.RestClient.parseResponse(RestClient.java:390)
        at 
org.apache.flink.runtime.rest.RestClient.lambda$submitRequest$3(RestClient.java:374)
        at 
java.util.concurrent.CompletableFuture.uniCompose(CompletableFuture.java:966)
        at 
java.util.concurrent.CompletableFuture$UniCompose.tryFire(CompletableFuture.java:940)
        ... 4 more
   ```
   version: 
   flink 1.11.3
   hive 3.1.2
   iceberg-flink-runtime-0.10.0
   
   description:
   I got this error using iceberg, I want to know what caused it.
   


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]



---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to