massdosage commented on pull request #1407:
URL: https://github.com/apache/iceberg/pull/1407#issuecomment-694186791
Hey @pvary, apologies for delays in looking at this. Today I tried this out
"for real" on a Hive cluster we have running on EMR where I've been doing my
InputFormat tests. I built a hive-runtime jar from this PR and added to the
classpath of the Hive shell. I then ran:
`CREATE EXTERNAL TABLE bdp.iceberg_write_table STORED BY
'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION
'hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/table_a';`
However this fails with the following exception that I see in the Hive logs:
```
2020-09-17T11:59:02,006 INFO [main([])]: conf.HiveConf
(HiveConf.java:getLogIdVar(3957)) - Using the default value passed in for log
id: 0e8d5f3d-e100-4e3b-bae1-f272b504b04a
2020-09-17T11:59:02,007 INFO [main([])]: session.SessionState
(SessionState.java:updateThreadName(421)) - Updating thread name to
0e8d5f3d-e100-4e3b-bae1-f272b504b04a main
2020-09-17T11:59:02,007 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:compile(429)) - Compiling
command(queryId=hadoop_20200917115902_126001a6-6193-4e7e-8def-345036e43056):
CREATE TABLE bdp.iceberg_write_table STORED BY
'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION
'hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table'
TBLPROPERTIES ('iceberg.mr.write.file.format'='parquet')
2020-09-17T11:59:02,010 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: parse.CalcitePlanner (SemanticAnalyzer.java:analyzeInternal(11150))
- Starting Semantic Analysis
2020-09-17T11:59:02,011 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: parse.CalcitePlanner
(SemanticAnalyzer.java:analyzeCreateTable(11896)) - Creating table
bdp.iceberg_write_table position=13
2020-09-17T11:59:02,043 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:compile(518)) - Semantic Analysis Completed
2020-09-17T11:59:02,044 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:getSchema(290)) - Returning Hive schema:
Schema(fieldSchemas:null, properties:null)
2020-09-17T11:59:02,044 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:compile(642)) - Completed compiling
command(queryId=hadoop_20200917115902_126001a6-6193-4e7e-8def-345036e43056);
Time taken: 0.037 seconds
2020-09-17T11:59:02,044 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:checkConcurrency(210)) - Concurrency mode is
disabled, not creating a lock manager
2020-09-17T11:59:02,044 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:execute(1735)) - Executing
command(queryId=hadoop_20200917115902_126001a6-6193-4e7e-8def-345036e43056):
CREATE TABLE bdp.iceberg_write_table STORED BY
'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' LOCATION
'hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table'
TBLPROPERTIES ('iceberg.mr.write.file.format'='parquet')
2020-09-17T11:59:02,091 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:launchTask(2181)) - Starting task
[Stage-0:DDL] in serial mode
2020-09-17T11:59:02,091 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: plan.CreateTableDesc (CreateTableDesc.java:toTable(717)) - Use
StorageHandler-supplied org.apache.iceberg.mr.hive.HiveIcebergSerDe for table
iceberg_write_table
2020-09-17T11:59:02,097 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: exec.DDLTask (DDLTask.java:createTable(4324)) - creating table
bdp.iceberg_write_table on
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table
2020-09-17T11:59:02,116 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: mr.Catalogs (Catalogs.java:loadCatalog(114)) - Catalog is not
configured
2020-09-17T11:59:02,124 WARN [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: hadoop.HadoopTableOperations
(HadoopTableOperations.java:readVersionHint(292)) - Error reading version hint
file
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table/metadata/version-hint.text
java.io.FileNotFoundException: File does not exist:
/hiveberg/iceberg_write_table/metadata/version-hint.text
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:152)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1827)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:692)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:381)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:850)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:793)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2489)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[?:1.8.0_212]
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[?:1.8.0_212]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
~[?:1.8.0_212]
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:849)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:836)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:825)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:325)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:286)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:270)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1064)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:329)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:325)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:337)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:790)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.readVersionHint(HadoopTableOperations.java:287)
~[?:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.refresh(HadoopTableOperations.java:94)
~[?:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.current(HadoopTableOperations.java:75)
~[?:?]
at org.apache.iceberg.hadoop.HadoopTables.load(HadoopTables.java:79)
~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:77) ~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:65) ~[?:?]
at
org.apache.iceberg.mr.hive.HiveIcebergSerDe.initialize(HiveIcebergSerDe.java:66)
~[?:?]
at
org.apache.hadoop.hive.serde2.AbstractSerDe.initialize(AbstractSerDe.java:54)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:533)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:449)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:436)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:281)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:263)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:641)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:624)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:836)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:872)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.8.0_212]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_212]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_212]
at org.apache.hadoop.util.RunJar.run(RunJar.java:239)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.util.RunJar.main(RunJar.java:153)
~[hadoop-common-2.8.5-amzn-4.jar:?]
Caused by: org.apache.hadoop.ipc.RemoteException: File does not exist:
/hiveberg/iceberg_write_table/metadata/version-hint.text
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:152)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1827)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:692)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:381)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:850)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:793)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2489)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.ipc.Client.call(Client.java:1435)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.ipc.Client.call(Client.java:1345)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at com.sun.proxy.$Proxy29.getBlockLocations(Unknown Source) ~[?:?]
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:259)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.8.0_212]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_212]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_212]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at com.sun.proxy.$Proxy30.getBlockLocations(Unknown Source) ~[?:?]
at
org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:847)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
... 49 more
2020-09-17T11:59:02,144 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: conf.HiveConf (HiveConf.java:getLogIdVar(3957)) - Using the default
value passed in for log id: 0e8d5f3d-e100-4e3b-bae1-f272b504b04a
2020-09-17T11:59:02,144 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: session.SessionState (SessionState.java:resetThreadName(432)) -
Resetting thread name to main
2020-09-17T11:59:02,128 ERROR [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: metadata.Table (Table.java:getColsInternal(644)) - Unable to get
field from serde: org.apache.iceberg.mr.hive.HiveIcebergSerDe
org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at
location:
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table
at org.apache.iceberg.hadoop.HadoopTables.load(HadoopTables.java:82)
~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:77) ~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:65) ~[?:?]
at
org.apache.iceberg.mr.hive.HiveIcebergSerDe.initialize(HiveIcebergSerDe.java:66)
~[?:?]
at
org.apache.hadoop.hive.serde2.AbstractSerDe.initialize(AbstractSerDe.java:54)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:533)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:449)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:436)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:281)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:263)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getColsInternal(Table.java:641)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.metadata.Table.getCols(Table.java:624)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:836)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:872)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.8.0_212]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_212]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_212]
at org.apache.hadoop.util.RunJar.run(RunJar.java:239)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.util.RunJar.main(RunJar.java:153)
~[hadoop-common-2.8.5-amzn-4.jar:?]
2020-09-17T11:59:02,128 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: mr.Catalogs (Catalogs.java:loadCatalog(114)) - Catalog is not
configured
2020-09-17T11:59:02,129 WARN [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: hadoop.HadoopTableOperations
(HadoopTableOperations.java:readVersionHint(292)) - Error reading version hint
file
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table/metadata/version-hint.text
java.io.FileNotFoundException: File does not exist:
/hiveberg/iceberg_write_table/metadata/version-hint.text
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:152)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1827)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:692)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:381)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:850)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:793)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2489)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
Method) ~[?:1.8.0_212]
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
~[?:1.8.0_212]
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
~[?:1.8.0_212]
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:121)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:88)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:849)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:836)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSClient.getLocatedBlocks(DFSClient.java:825)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.fetchLocatedBlocksAndGetLastBlockLength(DFSInputStream.java:325)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.openInfo(DFSInputStream.java:286)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DFSInputStream.<init>(DFSInputStream.java:270)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:1064)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:329)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem$4.doCall(DistributedFileSystem.java:325)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:337)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:790)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.readVersionHint(HadoopTableOperations.java:287)
~[?:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.refresh(HadoopTableOperations.java:94)
~[?:?]
at
org.apache.iceberg.hadoop.HadoopTableOperations.current(HadoopTableOperations.java:75)
~[?:?]
at org.apache.iceberg.hadoop.HadoopTables.load(HadoopTables.java:79)
~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:77) ~[?:?]
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:65) ~[?:?]
at
org.apache.iceberg.mr.hive.HiveIcebergSerDe.initialize(HiveIcebergSerDe.java:66)
~[?:?]
at
org.apache.hadoop.hive.serde2.AbstractSerDe.initialize(AbstractSerDe.java:54)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:533)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:449)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:436)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:281)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:263)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:838)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:872)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
~[hive-exec-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686)
~[hive-cli-2.3.5-amzn-0.jar:2.3.5-amzn-0]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.8.0_212]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_212]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_212]
at org.apache.hadoop.util.RunJar.run(RunJar.java:239)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.util.RunJar.main(RunJar.java:153)
~[hadoop-common-2.8.5-amzn-4.jar:?]
Caused by: org.apache.hadoop.ipc.RemoteException: File does not exist:
/hiveberg/iceberg_write_table/metadata/version-hint.text
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:72)
at
org.apache.hadoop.hdfs.server.namenode.INodeFile.valueOf(INodeFile.java:62)
at
org.apache.hadoop.hdfs.server.namenode.FSDirStatAndListingOp.getBlockLocations(FSDirStatAndListingOp.java:152)
at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getBlockLocations(FSNamesystem.java:1827)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.getBlockLocations(NameNodeRpcServer.java:692)
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.getBlockLocations(ClientNamenodeProtocolServerSideTranslatorPB.java:381)
at
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:447)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:989)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:850)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:793)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1844)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2489)
at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1489)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.ipc.Client.call(Client.java:1435)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at org.apache.hadoop.ipc.Client.call(Client.java:1345)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:227)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at com.sun.proxy.$Proxy29.getBlockLocations(Unknown Source) ~[?:?]
at
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getBlockLocations(ClientNamenodeProtocolTranslatorPB.java:259)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
~[?:1.8.0_212]
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
~[?:1.8.0_212]
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
~[?:1.8.0_212]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_212]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:409)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:163)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:155)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:346)
~[hadoop-common-2.8.5-amzn-4.jar:?]
at com.sun.proxy.$Proxy30.getBlockLocations(Unknown Source) ~[?:?]
at
org.apache.hadoop.hdfs.DFSClient.callGetBlockLocations(DFSClient.java:847)
~[hadoop-hdfs-client-2.8.5-amzn-4.jar:?]
... 47 more
2020-09-17T11:59:02,143 ERROR [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: exec.DDLTask (DDLTask.java:failed(639)) -
org.apache.hadoop.hive.ql.metadata.HiveException:
org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at
location:
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table
at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:867)
at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:872)
at
org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356)
at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354)
at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199)
at
org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100)
at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183)
at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839)
at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237)
at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1227)
at
org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:233)
at
org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:184)
at
org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:403)
at
org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:821)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:759)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:686)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.hadoop.util.RunJar.run(RunJar.java:239)
at org.apache.hadoop.util.RunJar.main(RunJar.java:153)
Caused by: org.apache.iceberg.exceptions.NoSuchTableException: Table does
not exist at location:
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table
at org.apache.iceberg.hadoop.HadoopTables.load(HadoopTables.java:82)
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:77)
at org.apache.iceberg.mr.Catalogs.loadTable(Catalogs.java:65)
at
org.apache.iceberg.mr.hive.HiveIcebergSerDe.initialize(HiveIcebergSerDe.java:66)
at
org.apache.hadoop.hive.serde2.AbstractSerDe.initialize(AbstractSerDe.java:54)
at
org.apache.hadoop.hive.serde2.SerDeUtils.initializeSerDe(SerDeUtils.java:533)
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:449)
at
org.apache.hadoop.hive.metastore.MetaStoreUtils.getDeserializer(MetaStoreUtils.java:436)
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializerFromMetaStore(Table.java:281)
at
org.apache.hadoop.hive.ql.metadata.Table.getDeserializer(Table.java:263)
at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:838)
... 22 more
2020-09-17T11:59:02,143 ERROR [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (SessionState.java:printError(1126)) - FAILED: Execution
Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask.
org.apache.iceberg.exceptions.NoSuchTableException: Table does not exist at
location:
hdfs://ip-redacted.us-west-2.compute.internal:8020/hiveberg/iceberg_write_table
2020-09-17T11:59:02,143 INFO [0e8d5f3d-e100-4e3b-bae1-f272b504b04a
main([])]: ql.Driver (Driver.java:execute(2050)) - Completed executing
command(queryId=hadoop_20200917115902_126001a6-6193-4e7e-8def-345036e43056);
Time taken: 0.099 seconds
```
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]