http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py index 923fe2c..7fc1e43 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-py/hive_metastore/ttypes.py @@ -2510,7 +2510,6 @@ class TruncateTableRequest: - dbName - tableName - partNames - - txnId - writeId - validWriteIdList """ @@ -2520,16 +2519,14 @@ class TruncateTableRequest: (1, TType.STRING, 'dbName', None, None, ), # 1 (2, TType.STRING, 'tableName', None, None, ), # 2 (3, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 3 - (4, TType.I64, 'txnId', None, -1, ), # 4 - (5, TType.I64, 'writeId', None, -1, ), # 5 - (6, TType.STRING, 'validWriteIdList', None, None, ), # 6 + (4, TType.I64, 'writeId', None, -1, ), # 4 + (5, TType.STRING, 'validWriteIdList', None, None, ), # 5 ) - def __init__(self, dbName=None, tableName=None, partNames=None, txnId=thrift_spec[4][4], writeId=thrift_spec[5][4], validWriteIdList=None,): + def __init__(self, dbName=None, tableName=None, partNames=None, writeId=thrift_spec[4][4], validWriteIdList=None,): self.dbName = dbName self.tableName = tableName self.partNames = partNames - self.txnId = txnId self.writeId = writeId self.validWriteIdList = validWriteIdList @@ -2564,15 +2561,10 @@ class TruncateTableRequest: iprot.skip(ftype) elif fid == 4: if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 5: - if ftype == TType.I64: self.writeId = iprot.readI64() else: iprot.skip(ftype) - elif fid == 6: + elif fid == 5: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -2602,16 +2594,12 @@ class TruncateTableRequest: oprot.writeString(iter75) oprot.writeListEnd() oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 4) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.writeId is not None: - oprot.writeFieldBegin('writeId', TType.I64, 5) + oprot.writeFieldBegin('writeId', TType.I64, 4) oprot.writeI64(self.writeId) oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -2630,7 +2618,6 @@ class TruncateTableRequest: value = (value * 31) ^ hash(self.dbName) value = (value * 31) ^ hash(self.tableName) value = (value * 31) ^ hash(self.partNames) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.writeId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -7316,7 +7303,6 @@ class SetPartitionsStatsRequest: Attributes: - colStats - needMerge - - txnId - writeId - validWriteIdList """ @@ -7325,15 +7311,13 @@ class SetPartitionsStatsRequest: None, # 0 (1, TType.LIST, 'colStats', (TType.STRUCT,(ColumnStatistics, ColumnStatistics.thrift_spec)), None, ), # 1 (2, TType.BOOL, 'needMerge', None, None, ), # 2 - (3, TType.I64, 'txnId', None, -1, ), # 3 - (4, TType.I64, 'writeId', None, -1, ), # 4 - (5, TType.STRING, 'validWriteIdList', None, None, ), # 5 + (3, TType.I64, 'writeId', None, -1, ), # 3 + (4, TType.STRING, 'validWriteIdList', None, None, ), # 4 ) - def __init__(self, colStats=None, needMerge=None, txnId=thrift_spec[3][4], writeId=thrift_spec[4][4], validWriteIdList=None,): + def __init__(self, colStats=None, needMerge=None, writeId=thrift_spec[3][4], validWriteIdList=None,): self.colStats = colStats self.needMerge = needMerge - self.txnId = txnId self.writeId = writeId self.validWriteIdList = validWriteIdList @@ -7364,15 +7348,10 @@ class SetPartitionsStatsRequest: iprot.skip(ftype) elif fid == 3: if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 4: - if ftype == TType.I64: self.writeId = iprot.readI64() else: iprot.skip(ftype) - elif fid == 5: + elif fid == 4: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -7398,16 +7377,12 @@ class SetPartitionsStatsRequest: oprot.writeFieldBegin('needMerge', TType.BOOL, 2) oprot.writeBool(self.needMerge) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 3) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.writeId is not None: - oprot.writeFieldBegin('writeId', TType.I64, 4) + oprot.writeFieldBegin('writeId', TType.I64, 3) oprot.writeI64(self.writeId) oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 4) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -7423,7 +7398,6 @@ class SetPartitionsStatsRequest: value = 17 value = (value * 31) ^ hash(self.colStats) value = (value * 31) ^ hash(self.needMerge) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.writeId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -9722,7 +9696,6 @@ class TableStatsRequest: - tblName - colNames - catName - - txnId - validWriteIdList """ @@ -9732,16 +9705,14 @@ class TableStatsRequest: (2, TType.STRING, 'tblName', None, None, ), # 2 (3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3 (4, TType.STRING, 'catName', None, None, ), # 4 - (5, TType.I64, 'txnId', None, -1, ), # 5 - (6, TType.STRING, 'validWriteIdList', None, None, ), # 6 + (5, TType.STRING, 'validWriteIdList', None, None, ), # 5 ) - def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,): + def __init__(self, dbName=None, tblName=None, colNames=None, catName=None, validWriteIdList=None,): self.dbName = dbName self.tblName = tblName self.colNames = colNames self.catName = catName - self.txnId = txnId self.validWriteIdList = validWriteIdList def read(self, iprot): @@ -9779,11 +9750,6 @@ class TableStatsRequest: else: iprot.skip(ftype) elif fid == 5: - if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 6: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -9817,12 +9783,8 @@ class TableStatsRequest: oprot.writeFieldBegin('catName', TType.STRING, 4) oprot.writeString(self.catName) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 5) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 5) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -9844,7 +9806,6 @@ class TableStatsRequest: value = (value * 31) ^ hash(self.tblName) value = (value * 31) ^ hash(self.colNames) value = (value * 31) ^ hash(self.catName) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -9867,7 +9828,6 @@ class PartitionsStatsRequest: - colNames - partNames - catName - - txnId - validWriteIdList """ @@ -9878,17 +9838,15 @@ class PartitionsStatsRequest: (3, TType.LIST, 'colNames', (TType.STRING,None), None, ), # 3 (4, TType.LIST, 'partNames', (TType.STRING,None), None, ), # 4 (5, TType.STRING, 'catName', None, None, ), # 5 - (6, TType.I64, 'txnId', None, -1, ), # 6 - (7, TType.STRING, 'validWriteIdList', None, None, ), # 7 + (6, TType.STRING, 'validWriteIdList', None, None, ), # 6 ) - def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, txnId=thrift_spec[6][4], validWriteIdList=None,): + def __init__(self, dbName=None, tblName=None, colNames=None, partNames=None, catName=None, validWriteIdList=None,): self.dbName = dbName self.tblName = tblName self.colNames = colNames self.partNames = partNames self.catName = catName - self.txnId = txnId self.validWriteIdList = validWriteIdList def read(self, iprot): @@ -9936,11 +9894,6 @@ class PartitionsStatsRequest: else: iprot.skip(ftype) elif fid == 6: - if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 7: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -9981,12 +9934,8 @@ class PartitionsStatsRequest: oprot.writeFieldBegin('catName', TType.STRING, 5) oprot.writeString(self.catName) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 6) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -10011,7 +9960,6 @@ class PartitionsStatsRequest: value = (value * 31) ^ hash(self.colNames) value = (value * 31) ^ hash(self.partNames) value = (value * 31) ^ hash(self.catName) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -10122,7 +10070,6 @@ class AddPartitionsRequest: - ifNotExists - needResult - catName - - txnId - validWriteIdList """ @@ -10134,18 +10081,16 @@ class AddPartitionsRequest: (4, TType.BOOL, 'ifNotExists', None, None, ), # 4 (5, TType.BOOL, 'needResult', None, True, ), # 5 (6, TType.STRING, 'catName', None, None, ), # 6 - (7, TType.I64, 'txnId', None, -1, ), # 7 - (8, TType.STRING, 'validWriteIdList', None, None, ), # 8 + (7, TType.STRING, 'validWriteIdList', None, None, ), # 7 ) - def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, txnId=thrift_spec[7][4], validWriteIdList=None,): + def __init__(self, dbName=None, tblName=None, parts=None, ifNotExists=None, needResult=thrift_spec[5][4], catName=None, validWriteIdList=None,): self.dbName = dbName self.tblName = tblName self.parts = parts self.ifNotExists = ifNotExists self.needResult = needResult self.catName = catName - self.txnId = txnId self.validWriteIdList = validWriteIdList def read(self, iprot): @@ -10194,11 +10139,6 @@ class AddPartitionsRequest: else: iprot.skip(ftype) elif fid == 7: - if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 8: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -10240,12 +10180,8 @@ class AddPartitionsRequest: oprot.writeFieldBegin('catName', TType.STRING, 6) oprot.writeString(self.catName) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 7) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -10271,7 +10207,6 @@ class AddPartitionsRequest: value = (value * 31) ^ hash(self.ifNotExists) value = (value * 31) ^ hash(self.needResult) value = (value * 31) ^ hash(self.catName) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -17526,7 +17461,6 @@ class GetTableRequest: - tblName - capabilities - catName - - txnId - validWriteIdList """ @@ -17536,16 +17470,15 @@ class GetTableRequest: (2, TType.STRING, 'tblName', None, None, ), # 2 (3, TType.STRUCT, 'capabilities', (ClientCapabilities, ClientCapabilities.thrift_spec), None, ), # 3 (4, TType.STRING, 'catName', None, None, ), # 4 - (5, TType.I64, 'txnId', None, -1, ), # 5 + None, # 5 (6, TType.STRING, 'validWriteIdList', None, None, ), # 6 ) - def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, txnId=thrift_spec[5][4], validWriteIdList=None,): + def __init__(self, dbName=None, tblName=None, capabilities=None, catName=None, validWriteIdList=None,): self.dbName = dbName self.tblName = tblName self.capabilities = capabilities self.catName = catName - self.txnId = txnId self.validWriteIdList = validWriteIdList def read(self, iprot): @@ -17578,11 +17511,6 @@ class GetTableRequest: self.catName = iprot.readString() else: iprot.skip(ftype) - elif fid == 5: - if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) elif fid == 6: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() @@ -17614,10 +17542,6 @@ class GetTableRequest: oprot.writeFieldBegin('catName', TType.STRING, 4) oprot.writeString(self.catName) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 5) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.validWriteIdList is not None: oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6) oprot.writeString(self.validWriteIdList) @@ -17639,7 +17563,6 @@ class GetTableRequest: value = (value * 31) ^ hash(self.tblName) value = (value * 31) ^ hash(self.capabilities) value = (value * 31) ^ hash(self.catName) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -22583,7 +22506,6 @@ class AlterPartitionsRequest: - tableName - partitions - environmentContext - - txnId - writeId - validWriteIdList """ @@ -22595,18 +22517,16 @@ class AlterPartitionsRequest: (3, TType.STRING, 'tableName', None, None, ), # 3 (4, TType.LIST, 'partitions', (TType.STRUCT,(Partition, Partition.thrift_spec)), None, ), # 4 (5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5 - (6, TType.I64, 'txnId', None, -1, ), # 6 - (7, TType.I64, 'writeId', None, -1, ), # 7 - (8, TType.STRING, 'validWriteIdList', None, None, ), # 8 + (6, TType.I64, 'writeId', None, -1, ), # 6 + (7, TType.STRING, 'validWriteIdList', None, None, ), # 7 ) - def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,): + def __init__(self, catName=None, dbName=None, tableName=None, partitions=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,): self.catName = catName self.dbName = dbName self.tableName = tableName self.partitions = partitions self.environmentContext = environmentContext - self.txnId = txnId self.writeId = writeId self.validWriteIdList = validWriteIdList @@ -22653,15 +22573,10 @@ class AlterPartitionsRequest: iprot.skip(ftype) elif fid == 6: if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 7: - if ftype == TType.I64: self.writeId = iprot.readI64() else: iprot.skip(ftype) - elif fid == 8: + elif fid == 7: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -22699,16 +22614,12 @@ class AlterPartitionsRequest: oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5) self.environmentContext.write(oprot) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 6) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.writeId is not None: - oprot.writeFieldBegin('writeId', TType.I64, 7) + oprot.writeFieldBegin('writeId', TType.I64, 6) oprot.writeI64(self.writeId) oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -22731,7 +22642,6 @@ class AlterPartitionsRequest: value = (value * 31) ^ hash(self.tableName) value = (value * 31) ^ hash(self.partitions) value = (value * 31) ^ hash(self.environmentContext) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.writeId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -22801,7 +22711,6 @@ class RenamePartitionRequest: - tableName - partVals - newPart - - txnId - validWriteIdList """ @@ -22812,17 +22721,15 @@ class RenamePartitionRequest: (3, TType.STRING, 'tableName', None, None, ), # 3 (4, TType.LIST, 'partVals', (TType.STRING,None), None, ), # 4 (5, TType.STRUCT, 'newPart', (Partition, Partition.thrift_spec), None, ), # 5 - (6, TType.I64, 'txnId', None, -1, ), # 6 - (7, TType.STRING, 'validWriteIdList', None, None, ), # 7 + (6, TType.STRING, 'validWriteIdList', None, None, ), # 6 ) - def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, txnId=thrift_spec[6][4], validWriteIdList=None,): + def __init__(self, catName=None, dbName=None, tableName=None, partVals=None, newPart=None, validWriteIdList=None,): self.catName = catName self.dbName = dbName self.tableName = tableName self.partVals = partVals self.newPart = newPart - self.txnId = txnId self.validWriteIdList = validWriteIdList def read(self, iprot): @@ -22866,11 +22773,6 @@ class RenamePartitionRequest: else: iprot.skip(ftype) elif fid == 6: - if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 7: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -22908,12 +22810,8 @@ class RenamePartitionRequest: oprot.writeFieldBegin('newPart', TType.STRUCT, 5) self.newPart.write(oprot) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 6) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 6) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -22938,7 +22836,6 @@ class RenamePartitionRequest: value = (value * 31) ^ hash(self.tableName) value = (value * 31) ^ hash(self.partVals) value = (value * 31) ^ hash(self.newPart) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.validWriteIdList) return value @@ -23007,7 +22904,6 @@ class AlterTableRequest: - tableName - table - environmentContext - - txnId - writeId - validWriteIdList """ @@ -23019,18 +22915,16 @@ class AlterTableRequest: (3, TType.STRING, 'tableName', None, None, ), # 3 (4, TType.STRUCT, 'table', (Table, Table.thrift_spec), None, ), # 4 (5, TType.STRUCT, 'environmentContext', (EnvironmentContext, EnvironmentContext.thrift_spec), None, ), # 5 - (6, TType.I64, 'txnId', None, -1, ), # 6 - (7, TType.I64, 'writeId', None, -1, ), # 7 - (8, TType.STRING, 'validWriteIdList', None, None, ), # 8 + (6, TType.I64, 'writeId', None, -1, ), # 6 + (7, TType.STRING, 'validWriteIdList', None, None, ), # 7 ) - def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, txnId=thrift_spec[6][4], writeId=thrift_spec[7][4], validWriteIdList=None,): + def __init__(self, catName=None, dbName=None, tableName=None, table=None, environmentContext=None, writeId=thrift_spec[6][4], validWriteIdList=None,): self.catName = catName self.dbName = dbName self.tableName = tableName self.table = table self.environmentContext = environmentContext - self.txnId = txnId self.writeId = writeId self.validWriteIdList = validWriteIdList @@ -23072,15 +22966,10 @@ class AlterTableRequest: iprot.skip(ftype) elif fid == 6: if ftype == TType.I64: - self.txnId = iprot.readI64() - else: - iprot.skip(ftype) - elif fid == 7: - if ftype == TType.I64: self.writeId = iprot.readI64() else: iprot.skip(ftype) - elif fid == 8: + elif fid == 7: if ftype == TType.STRING: self.validWriteIdList = iprot.readString() else: @@ -23115,16 +23004,12 @@ class AlterTableRequest: oprot.writeFieldBegin('environmentContext', TType.STRUCT, 5) self.environmentContext.write(oprot) oprot.writeFieldEnd() - if self.txnId is not None: - oprot.writeFieldBegin('txnId', TType.I64, 6) - oprot.writeI64(self.txnId) - oprot.writeFieldEnd() if self.writeId is not None: - oprot.writeFieldBegin('writeId', TType.I64, 7) + oprot.writeFieldBegin('writeId', TType.I64, 6) oprot.writeI64(self.writeId) oprot.writeFieldEnd() if self.validWriteIdList is not None: - oprot.writeFieldBegin('validWriteIdList', TType.STRING, 8) + oprot.writeFieldBegin('validWriteIdList', TType.STRING, 7) oprot.writeString(self.validWriteIdList) oprot.writeFieldEnd() oprot.writeFieldStop() @@ -23147,7 +23032,6 @@ class AlterTableRequest: value = (value * 31) ^ hash(self.tableName) value = (value * 31) ^ hash(self.table) value = (value * 31) ^ hash(self.environmentContext) - value = (value * 31) ^ hash(self.txnId) value = (value * 31) ^ hash(self.writeId) value = (value * 31) ^ hash(self.validWriteIdList) return value
http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb index 2e2392d..e0c6c02 100644 --- a/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb +++ b/standalone-metastore/metastore-common/src/gen/thrift/gen-rb/hive_metastore_types.rb @@ -608,15 +608,13 @@ class TruncateTableRequest DBNAME = 1 TABLENAME = 2 PARTNAMES = 3 - TXNID = 4 - WRITEID = 5 - VALIDWRITEIDLIST = 6 + WRITEID = 4 + VALIDWRITEIDLIST = 5 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'}, TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}, :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -1646,14 +1644,12 @@ class SetPartitionsStatsRequest include ::Thrift::Struct, ::Thrift::Struct_Union COLSTATS = 1 NEEDMERGE = 2 - TXNID = 3 - WRITEID = 4 - VALIDWRITEIDLIST = 5 + WRITEID = 3 + VALIDWRITEIDLIST = 4 FIELDS = { COLSTATS => {:type => ::Thrift::Types::LIST, :name => 'colStats', :element => {:type => ::Thrift::Types::STRUCT, :class => ::ColumnStatistics}}, NEEDMERGE => {:type => ::Thrift::Types::BOOL, :name => 'needMerge', :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -2178,15 +2174,13 @@ class TableStatsRequest TBLNAME = 2 COLNAMES = 3 CATNAME = 4 - TXNID = 5 - VALIDWRITEIDLIST = 6 + VALIDWRITEIDLIST = 5 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'}, TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'}, COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}}, CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -2208,8 +2202,7 @@ class PartitionsStatsRequest COLNAMES = 3 PARTNAMES = 4 CATNAME = 5 - TXNID = 6 - VALIDWRITEIDLIST = 7 + VALIDWRITEIDLIST = 6 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'}, @@ -2217,7 +2210,6 @@ class PartitionsStatsRequest COLNAMES => {:type => ::Thrift::Types::LIST, :name => 'colNames', :element => {:type => ::Thrift::Types::STRING}}, PARTNAMES => {:type => ::Thrift::Types::LIST, :name => 'partNames', :element => {:type => ::Thrift::Types::STRING}}, CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -2259,8 +2251,7 @@ class AddPartitionsRequest IFNOTEXISTS = 4 NEEDRESULT = 5 CATNAME = 6 - TXNID = 7 - VALIDWRITEIDLIST = 8 + VALIDWRITEIDLIST = 7 FIELDS = { DBNAME => {:type => ::Thrift::Types::STRING, :name => 'dbName'}, @@ -2269,7 +2260,6 @@ class AddPartitionsRequest IFNOTEXISTS => {:type => ::Thrift::Types::BOOL, :name => 'ifNotExists'}, NEEDRESULT => {:type => ::Thrift::Types::BOOL, :name => 'needResult', :default => true, :optional => true}, CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -3915,7 +3905,6 @@ class GetTableRequest TBLNAME = 2 CAPABILITIES = 3 CATNAME = 4 - TXNID = 5 VALIDWRITEIDLIST = 6 FIELDS = { @@ -3923,7 +3912,6 @@ class GetTableRequest TBLNAME => {:type => ::Thrift::Types::STRING, :name => 'tblName'}, CAPABILITIES => {:type => ::Thrift::Types::STRUCT, :name => 'capabilities', :class => ::ClientCapabilities, :optional => true}, CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -5114,9 +5102,8 @@ class AlterPartitionsRequest TABLENAME = 3 PARTITIONS = 4 ENVIRONMENTCONTEXT = 5 - TXNID = 6 - WRITEID = 7 - VALIDWRITEIDLIST = 8 + WRITEID = 6 + VALIDWRITEIDLIST = 7 FIELDS = { CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, @@ -5124,7 +5111,6 @@ class AlterPartitionsRequest TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, PARTITIONS => {:type => ::Thrift::Types::LIST, :name => 'partitions', :element => {:type => ::Thrift::Types::STRUCT, :class => ::Partition}}, ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -5162,8 +5148,7 @@ class RenamePartitionRequest TABLENAME = 3 PARTVALS = 4 NEWPART = 5 - TXNID = 6 - VALIDWRITEIDLIST = 7 + VALIDWRITEIDLIST = 6 FIELDS = { CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, @@ -5171,7 +5156,6 @@ class RenamePartitionRequest TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, PARTVALS => {:type => ::Thrift::Types::LIST, :name => 'partVals', :element => {:type => ::Thrift::Types::STRING}}, NEWPART => {:type => ::Thrift::Types::STRUCT, :name => 'newPart', :class => ::Partition}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } @@ -5209,9 +5193,8 @@ class AlterTableRequest TABLENAME = 3 TABLE = 4 ENVIRONMENTCONTEXT = 5 - TXNID = 6 - WRITEID = 7 - VALIDWRITEIDLIST = 8 + WRITEID = 6 + VALIDWRITEIDLIST = 7 FIELDS = { CATNAME => {:type => ::Thrift::Types::STRING, :name => 'catName', :optional => true}, @@ -5219,7 +5202,6 @@ class AlterTableRequest TABLENAME => {:type => ::Thrift::Types::STRING, :name => 'tableName'}, TABLE => {:type => ::Thrift::Types::STRUCT, :name => 'table', :class => ::Table}, ENVIRONMENTCONTEXT => {:type => ::Thrift::Types::STRUCT, :name => 'environmentContext', :class => ::EnvironmentContext, :optional => true}, - TXNID => {:type => ::Thrift::Types::I64, :name => 'txnId', :default => -1, :optional => true}, WRITEID => {:type => ::Thrift::Types::I64, :name => 'writeId', :default => -1, :optional => true}, VALIDWRITEIDLIST => {:type => ::Thrift::Types::STRING, :name => 'validWriteIdList', :optional => true} } http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift index 4cfa6a1..8a4bdd8 100644 --- a/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift +++ b/standalone-metastore/metastore-common/src/main/thrift/hive_metastore.thrift @@ -282,9 +282,8 @@ struct TruncateTableRequest { 1: required string dbName, 2: required string tableName, 3: optional list<string> partNames, - 4: optional i64 txnId=-1, - 5: optional i64 writeId=-1, - 6: optional string validWriteIdList + 4: optional i64 writeId=-1, + 5: optional string validWriteIdList } struct TruncateTableResponse { @@ -596,9 +595,8 @@ struct AggrStats { struct SetPartitionsStatsRequest { 1: required list<ColumnStatistics> colStats, 2: optional bool needMerge, //stats need to be merged with the existing stats -3: optional i64 txnId=-1, // transaction id of the query that sends this structure -4: optional i64 writeId=-1, // writeId for the current query that updates the stats -5: optional string validWriteIdList // valid write id list for the table for which this struct is being sent +3: optional i64 writeId=-1, // writeId for the current query that updates the stats +4: optional string validWriteIdList // valid write id list for the table for which this struct is being sent } struct SetPartitionsStatsResponse { @@ -745,8 +743,7 @@ struct TableStatsRequest { 2: required string tblName, 3: required list<string> colNames 4: optional string catName, - 5: optional i64 txnId=-1, // transaction id of the query that sends this structure - 6: optional string validWriteIdList // valid write id list for the table for which this struct is being sent + 5: optional string validWriteIdList // valid write id list for the table for which this struct is being sent } struct PartitionsStatsRequest { @@ -755,8 +752,7 @@ struct PartitionsStatsRequest { 3: required list<string> colNames, 4: required list<string> partNames, 5: optional string catName, - 6: optional i64 txnId=-1, // transaction id of the query that sends this structure - 7: optional string validWriteIdList // valid write id list for the table for which this struct is being sent + 6: optional string validWriteIdList // valid write id list for the table for which this struct is being sent } // Return type for add_partitions_req @@ -773,8 +769,7 @@ struct AddPartitionsRequest { 4: required bool ifNotExists, 5: optional bool needResult=true, 6: optional string catName, - 7: optional i64 txnId=-1, - 8: optional string validWriteIdList + 7: optional string validWriteIdList } // Return type for drop_partitions_req @@ -1275,7 +1270,6 @@ struct GetTableRequest { 2: required string tblName, 3: optional ClientCapabilities capabilities, 4: optional string catName, - 5: optional i64 txnId=-1, 6: optional string validWriteIdList } @@ -1615,9 +1609,8 @@ struct AlterPartitionsRequest { 3: required string tableName, 4: required list<Partition> partitions, 5: optional EnvironmentContext environmentContext, - 6: optional i64 txnId=-1, - 7: optional i64 writeId=-1, - 8: optional string validWriteIdList + 6: optional i64 writeId=-1, + 7: optional string validWriteIdList } struct AlterPartitionsResponse { @@ -1629,8 +1622,7 @@ struct RenamePartitionRequest { 3: required string tableName, 4: required list<string> partVals, 5: required Partition newPart, - 6: optional i64 txnId=-1, - 7: optional string validWriteIdList + 6: optional string validWriteIdList } struct RenamePartitionResponse { @@ -1642,9 +1634,8 @@ struct AlterTableRequest { 3: required string tableName, 4: required Table table, 5: optional EnvironmentContext environmentContext, - 6: optional i64 txnId=-1, - 7: optional i64 writeId=-1, - 8: optional string validWriteIdList + 6: optional i64 writeId=-1, + 7: optional string validWriteIdList // TODO: also add cascade here, out of envCtx } http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java index f3dc264..f4e0c41 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/AlterHandler.java @@ -61,7 +61,7 @@ public interface AlterHandler extends Configurable { default void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname, String name, Table newTable, EnvironmentContext envContext) throws InvalidOperationException, MetaException { - alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, -1, null); + alterTable(msdb, wh, catName, dbname, name, newTable, envContext, null, null); } /** @@ -88,7 +88,7 @@ public interface AlterHandler extends Configurable { */ void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname, String name, Table newTable, EnvironmentContext envContext, - IHMSHandler handler, long txnId, String writeIdList) + IHMSHandler handler, String writeIdList) throws InvalidOperationException, MetaException; /** @@ -146,7 +146,7 @@ public interface AlterHandler extends Configurable { Partition alterPartition(final RawStore msdb, Warehouse wh, final String catName, final String dbname, final String name, final List<String> part_vals, final Partition new_part, EnvironmentContext environmentContext, - IHMSHandler handler, long txnId, String validWriteIds) + IHMSHandler handler, String validWriteIds) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException; /** @@ -198,7 +198,7 @@ public interface AlterHandler extends Configurable { */ List<Partition> alterPartitions(final RawStore msdb, Warehouse wh, final String catName, final String dbname, final String name, final List<Partition> new_parts, - EnvironmentContext environmentContext, long txnId, String writeIdList, long writeId, + EnvironmentContext environmentContext, String writeIdList, long writeId, IHMSHandler handler) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException; } \ No newline at end of file http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java index 719f001..e16bf7a 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveAlterHandler.java @@ -90,7 +90,7 @@ public class HiveAlterHandler implements AlterHandler { @Override public void alterTable(RawStore msdb, Warehouse wh, String catName, String dbname, String name, Table newt, EnvironmentContext environmentContext, - IHMSHandler handler, long txnId, String writeIdList) + IHMSHandler handler, String writeIdList) throws InvalidOperationException, MetaException { catName = normalizeIdentifier(catName); name = name.toLowerCase(); @@ -142,7 +142,7 @@ public class HiveAlterHandler implements AlterHandler { // check if table with the new name already exists if (!newTblName.equals(name) || !newDbName.equals(dbname)) { - if (msdb.getTable(catName, newDbName, newTblName, -1, null) != null) { + if (msdb.getTable(catName, newDbName, newTblName, null) != null) { throw new InvalidOperationException("new table " + newDbName + "." + newTblName + " already exists"); } @@ -152,7 +152,7 @@ public class HiveAlterHandler implements AlterHandler { msdb.openTransaction(); // get old table // Note: we don't verify stats here; it's done below in alterTableUpdateTableColumnStats. - oldt = msdb.getTable(catName, dbname, name, -1, null); + oldt = msdb.getTable(catName, dbname, name, null); if (oldt == null) { throw new InvalidOperationException("table " + TableName.getQualified(catName, dbname, name) + " doesn't exist"); @@ -278,7 +278,7 @@ public class HiveAlterHandler implements AlterHandler { } } // Do not verify stats parameters on a partitioned table. - msdb.alterTable(catName, dbname, name, newt, -1, null); + msdb.alterTable(catName, dbname, name, newt, null); // alterPartition is only for changing the partition location in the table rename if (dataWasMoved) { @@ -297,7 +297,7 @@ public class HiveAlterHandler implements AlterHandler { partValues.add(part.getValues()); } msdb.alterPartitions(catName, newDbName, newTblName, partValues, - partBatch, newt.getWriteId(), txnId, writeIdList); + partBatch, newt.getWriteId(), writeIdList); } } @@ -306,11 +306,11 @@ public class HiveAlterHandler implements AlterHandler { newPartColStats.getStatsDesc().setDbName(newDbName); newPartColStats.getStatsDesc().setTableName(newTblName); msdb.updatePartitionColumnStatistics(newPartColStats, partColStats.getKey().getValues(), - txnId, writeIdList, newt.getWriteId()); + writeIdList, newt.getWriteId()); } } else { alterTableUpdateTableColumnStats( - msdb, oldt, newt, environmentContext, txnId, writeIdList); + msdb, oldt, newt, environmentContext, writeIdList); } } else { // operations other than table rename @@ -334,24 +334,24 @@ public class HiveAlterHandler implements AlterHandler { assert(colStats == null); if (cascade) { msdb.alterPartition( - catName, dbname, name, part.getValues(), part, txnId, writeIdList); + catName, dbname, name, part.getValues(), part, writeIdList); } else { // update changed properties (stats) oldPart.setParameters(part.getParameters()); msdb.alterPartition( - catName, dbname, name, part.getValues(), oldPart, txnId, writeIdList); + catName, dbname, name, part.getValues(), oldPart, writeIdList); } } // Don't validate table-level stats for a partitoned table. - msdb.alterTable(catName, dbname, name, newt, -1, null); + msdb.alterTable(catName, dbname, name, newt, null); } else { LOG.warn("Alter table not cascaded to partitions."); alterTableUpdateTableColumnStats( - msdb, oldt, newt, environmentContext, txnId, writeIdList); + msdb, oldt, newt, environmentContext, writeIdList); } } else { alterTableUpdateTableColumnStats( - msdb, oldt, newt, environmentContext, txnId, writeIdList); + msdb, oldt, newt, environmentContext, writeIdList); } } @@ -431,13 +431,13 @@ public class HiveAlterHandler implements AlterHandler { EnvironmentContext environmentContext) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException { return alterPartition(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, part_vals, new_part, - environmentContext, null, -1, null); + environmentContext, null, null); } @Override public Partition alterPartition(RawStore msdb, Warehouse wh, String catName, String dbname, String name, List<String> part_vals, final Partition new_part, - EnvironmentContext environmentContext, IHMSHandler handler, long txnId, String validWriteIds) + EnvironmentContext environmentContext, IHMSHandler handler, String validWriteIds) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException { boolean success = false; Partition oldPart; @@ -459,7 +459,7 @@ public class HiveAlterHandler implements AlterHandler { try { msdb.openTransaction(); - Table tbl = msdb.getTable(catName, dbname, name, -1, null); + Table tbl = msdb.getTable(catName, dbname, name, null); if (tbl == null) { throw new InvalidObjectException( "Unable to alter partition because table or database does not exist."); @@ -481,7 +481,7 @@ public class HiveAlterHandler implements AlterHandler { oldPart.getSd().getCols(), tbl, new_part, null); } msdb.alterPartition( - catName, dbname, name, new_part.getValues(), new_part, txnId, validWriteIds); + catName, dbname, name, new_part.getValues(), new_part, validWriteIds); if (transactionalListeners != null && !transactionalListeners.isEmpty()) { MetaStoreListenerNotifier.notifyEvent(transactionalListeners, EventMessage.EventType.ALTER_PARTITION, @@ -516,7 +516,7 @@ public class HiveAlterHandler implements AlterHandler { Database db; try { msdb.openTransaction(); - Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name, -1, null); + Table tbl = msdb.getTable(DEFAULT_CATALOG_NAME, dbname, name, null); if (tbl == null) { throw new InvalidObjectException( "Unable to alter partition because table or database does not exist."); @@ -617,12 +617,12 @@ public class HiveAlterHandler implements AlterHandler { String newPartName = Warehouse.makePartName(tbl.getPartitionKeys(), new_part.getValues()); ColumnStatistics cs = updateOrGetPartitionColumnStats(msdb, catName, dbname, name, oldPart.getValues(), oldPart.getSd().getCols(), tbl, new_part, null); - msdb.alterPartition(catName, dbname, name, part_vals, new_part, txnId, validWriteIds); + msdb.alterPartition(catName, dbname, name, part_vals, new_part, validWriteIds); if (cs != null) { cs.getStatsDesc().setPartName(newPartName); try { msdb.updatePartitionColumnStatistics(cs, new_part.getValues(), - txnId, validWriteIds, new_part.getWriteId()); + validWriteIds, new_part.getWriteId()); } catch (InvalidInputException iie) { throw new InvalidOperationException("Unable to update partition stats in table rename." + iie); } catch (NoSuchObjectException nsoe) { @@ -668,7 +668,7 @@ public class HiveAlterHandler implements AlterHandler { EnvironmentContext environmentContext) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException { return alterPartitions(msdb, wh, DEFAULT_CATALOG_NAME, dbname, name, new_parts, - environmentContext, -1, null, -1, null); + environmentContext, null, -1, null); } @Override @@ -676,7 +676,7 @@ public class HiveAlterHandler implements AlterHandler { final String dbname, final String name, final List<Partition> new_parts, EnvironmentContext environmentContext, - long txnId, String writeIdList, long writeId, + String writeIdList, long writeId, IHMSHandler handler) throws InvalidOperationException, InvalidObjectException, AlreadyExistsException, MetaException { List<Partition> oldParts = new ArrayList<>(); @@ -691,7 +691,7 @@ public class HiveAlterHandler implements AlterHandler { msdb.openTransaction(); // Note: should we pass in write ID here? We only update stats on parts so probably not. - Table tbl = msdb.getTable(catName, dbname, name, -1, null); + Table tbl = msdb.getTable(catName, dbname, name, null); if (tbl == null) { throw new InvalidObjectException( "Unable to alter partitions because table or database does not exist."); @@ -726,7 +726,7 @@ public class HiveAlterHandler implements AlterHandler { } } - msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, txnId, writeIdList); + msdb.alterPartitions(catName, dbname, name, partValsList, new_parts, writeId, writeIdList); Iterator<Partition> oldPartsIt = oldParts.iterator(); for (Partition newPart : new_parts) { Partition oldPart; @@ -795,7 +795,7 @@ public class HiveAlterHandler implements AlterHandler { @VisibleForTesting void alterTableUpdateTableColumnStats(RawStore msdb, Table oldTable, Table newTable, - EnvironmentContext ec, long txnId, String validWriteIds) + EnvironmentContext ec, String validWriteIds) throws MetaException, InvalidObjectException { String catName = normalizeIdentifier(oldTable.isSetCatName() ? oldTable.getCatName() : getDefaultCatalog(conf)); @@ -854,13 +854,13 @@ public class HiveAlterHandler implements AlterHandler { } // Change to new table and append stats for the new table - msdb.alterTable(catName, dbName, tableName, newTable, txnId, validWriteIds); + msdb.alterTable(catName, dbName, tableName, newTable, validWriteIds); if (updateColumnStats && !newStatsObjs.isEmpty()) { ColumnStatisticsDesc statsDesc = colStats.getStatsDesc(); statsDesc.setDbName(newDbName); statsDesc.setTableName(newTableName); colStats.setStatsObj(newStatsObjs); - msdb.updateTableColumnStatistics(colStats, txnId, validWriteIds, newTable.getWriteId()); + msdb.updateTableColumnStatistics(colStats, validWriteIds, newTable.getWriteId()); } } catch (NoSuchObjectException nsoe) { LOG.debug("Could not find db entry." + nsoe); http://git-wip-us.apache.org/repos/asf/hive/blob/cdb32a7f/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java ---------------------------------------------------------------------- diff --git a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java index 285f7fb..eb4eb1b 100644 --- a/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java +++ b/standalone-metastore/metastore-server/src/main/java/org/apache/hadoop/hive/metastore/HiveMetaStore.java @@ -2435,7 +2435,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { private boolean is_table_exists(RawStore ms, String catName, String dbname, String name) throws MetaException { - return (ms.getTable(catName, dbname, name, -1, null) != null); + return (ms.getTable(catName, dbname, name, null) != null); } private boolean drop_table_core(final RawStore ms, final String catName, final String dbname, @@ -2694,7 +2694,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } private void alterPartitionForTruncate(RawStore ms, String catName, String dbName, String tableName, - Table table, Partition partition, long txnId, String validWriteIds, long writeId) throws Exception { + Table table, Partition partition, String validWriteIds, long writeId) throws Exception { EnvironmentContext environmentContext = new EnvironmentContext(); updateStatsForTruncate(partition.getParameters(), environmentContext); @@ -2714,17 +2714,17 @@ public class HiveMetaStore extends ThriftHiveMetastore { partition.setWriteId(writeId); } alterHandler.alterPartition(ms, wh, catName, dbName, tableName, null, partition, - environmentContext, this, txnId, validWriteIds); + environmentContext, this, validWriteIds); } private void alterTableStatsForTruncate(RawStore ms, String catName, String dbName, String tableName, Table table, List<String> partNames, - long txnId, String validWriteIds, long writeId) throws Exception { + String validWriteIds, long writeId) throws Exception { if (partNames == null) { if (0 != table.getPartitionKeysSize()) { for (Partition partition : ms.getPartitions(catName, dbName, tableName, Integer.MAX_VALUE)) { alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition, - txnId, validWriteIds, writeId); + validWriteIds, writeId); } } else { EnvironmentContext environmentContext = new EnvironmentContext(); @@ -2747,12 +2747,12 @@ public class HiveMetaStore extends ThriftHiveMetastore { table.setWriteId(writeId); } alterHandler.alterTable(ms, wh, catName, dbName, tableName, table, - environmentContext, this, txnId, validWriteIds); + environmentContext, this, validWriteIds); } } else { for (Partition partition : ms.getPartitionsByNames(catName, dbName, tableName, partNames)) { alterPartitionForTruncate(ms, catName, dbName, tableName, table, partition, - txnId, validWriteIds, writeId); + validWriteIds, writeId); } } return; @@ -2791,19 +2791,19 @@ public class HiveMetaStore extends ThriftHiveMetastore { public void truncate_table(final String dbName, final String tableName, List<String> partNames) throws NoSuchObjectException, MetaException { // Deprecated path, won't work for txn tables. - truncateTableInternal(dbName, tableName, partNames, -1, null, -1); + truncateTableInternal(dbName, tableName, partNames, null, -1); } @Override public TruncateTableResponse truncate_table_req(TruncateTableRequest req) throws MetaException, TException { truncateTableInternal(req.getDbName(), req.getTableName(), req.getPartNames(), - req.getTxnId(), req.getValidWriteIdList(), req.getWriteId()); + req.getValidWriteIdList(), req.getWriteId()); return new TruncateTableResponse(); } private void truncateTableInternal(String dbName, String tableName, List<String> partNames, - long txnId, String validWriteIds, long writeId) throws MetaException, NoSuchObjectException { + String validWriteIds, long writeId) throws MetaException, NoSuchObjectException { try { String[] parsedDbName = parseDbName(dbName, conf); Table tbl = get_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName); @@ -2835,7 +2835,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { // Alter the table/partition stats and also notify truncate table event alterTableStatsForTruncate(getMS(), parsedDbName[CAT_NAME], parsedDbName[DB_NAME], - tableName, tbl, partNames, txnId, validWriteIds, writeId); + tableName, tbl, partNames, validWriteIds, writeId); } catch (IOException e) { throw new MetaException(e.getMessage()); } catch (MetaException | NoSuchObjectException e) { @@ -2866,7 +2866,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { NoSuchObjectException { String[] parsedDbName = parseDbName(dbname, conf); return getTableInternal( - parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, -1, null); + parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, null, null); } @Override @@ -2874,11 +2874,11 @@ public class HiveMetaStore extends ThriftHiveMetastore { NoSuchObjectException { String catName = req.isSetCatName() ? req.getCatName() : getDefaultCatalog(conf); return new GetTableResult(getTableInternal(catName, req.getDbName(), req.getTblName(), - req.getCapabilities(), req.getTxnId(), req.getValidWriteIdList())); + req.getCapabilities(), req.getValidWriteIdList())); } private Table getTableInternal(String catName, String dbname, String name, - ClientCapabilities capabilities, long txnId, String writeIdList) + ClientCapabilities capabilities, String writeIdList) throws MetaException, NoSuchObjectException { if (isInTest) { assertClientHasCapability(capabilities, ClientCapability.TEST_CAPABILITY, @@ -2889,7 +2889,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { startTableFunction("get_table", catName, dbname, name); Exception ex = null; try { - t = get_table_core(catName, dbname, name, txnId, writeIdList); + t = get_table_core(catName, dbname, name, writeIdList); if (MetaStoreUtils.isInsertOnlyTableParam(t.getParameters())) { assertClientHasCapability(capabilities, ClientCapability.INSERT_ONLY_TABLES, "insert-only tables", "get_table_req"); @@ -2929,7 +2929,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { final String dbname, final String name) throws MetaException, NoSuchObjectException { - return get_table_core(catName, dbname, name, -1, null); + return get_table_core(catName, dbname, name, null); } @Override @@ -2937,12 +2937,11 @@ public class HiveMetaStore extends ThriftHiveMetastore { final String catName, final String dbname, final String name, - final long txnId, final String writeIdList) throws MetaException, NoSuchObjectException { Table t = null; try { - t = getMS().getTable(catName, dbname, name, txnId, writeIdList); + t = getMS().getTable(catName, dbname, name, writeIdList); if (t == null) { throw new NoSuchObjectException(TableName.getQualified(catName, dbname, name) + " table not found"); @@ -3124,7 +3123,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { MetaStoreUtils.validatePartitionNameCharacters(part_vals, partitionValidationPattern); - tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null); + tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null); if (tbl == null) { throw new InvalidObjectException( "Unable to add partition because table or database do not exist"); @@ -3318,7 +3317,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { ms.openTransaction(); - tbl = ms.getTable(catName, dbName, tblName, -1, null); + tbl = ms.getTable(catName, dbName, tblName, null); if (tbl == null) { throw new InvalidObjectException("Unable to add partitions because " + TableName.getQualified(catName, dbName, tblName) + @@ -3674,7 +3673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { Database db = null; try { ms.openTransaction(); - tbl = ms.getTable(catName, dbName, tblName, -1, null); + tbl = ms.getTable(catName, dbName, tblName, null); if (tbl == null) { throw new InvalidObjectException("Unable to add partitions because " + "database or table " + dbName + "." + tblName + " does not exist"); @@ -3844,7 +3843,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } try { ms.openTransaction(); - tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), -1, null); + tbl = ms.getTable(part.getCatName(), part.getDbName(), part.getTableName(), null); if (tbl == null) { throw new InvalidObjectException( "Unable to add partition because table or database do not exist"); @@ -3966,7 +3965,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { Table destinationTable = ms.getTable( - parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, -1, null); + parsedDestDbName[CAT_NAME], parsedDestDbName[DB_NAME], destTableName, null); if (destinationTable == null) { throw new MetaException( "The destination table " + TableName.getQualified(parsedDestDbName[CAT_NAME], @@ -3974,7 +3973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } Table sourceTable = ms.getTable( - parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, -1, null); + parsedSourceDbName[CAT_NAME], parsedSourceDbName[DB_NAME], sourceTableName, null); if (sourceTable == null) { throw new MetaException("The source table " + TableName.getQualified(parsedSourceDbName[CAT_NAME], @@ -4151,7 +4150,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { ms.openTransaction(); part = ms.getPartition(catName, db_name, tbl_name, part_vals); - tbl = get_table_core(catName, db_name, tbl_name, -1, null); + tbl = get_table_core(catName, db_name, tbl_name, null); tableDataShouldBeDeleted = checkTableDataShouldBeDeleted(tbl, deleteData); firePreEvent(new PreDropPartitionEvent(tbl, part, deleteData, this)); mustPurge = isMustPurge(envContext, tbl); @@ -4826,7 +4825,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { String[] parsedDbName = parseDbName(dbName, conf); // TODO: this method name is confusing, it actually does full alter (sortof) rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, null, newPartition, - envContext, -1, null); + envContext, null); } @Deprecated @@ -4837,18 +4836,18 @@ public class HiveMetaStore extends ThriftHiveMetastore { // Call rename_partition without an environment context. String[] parsedDbName = parseDbName(db_name, conf); rename_partition(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tbl_name, part_vals, new_part, - null, -1, null); + null, null); } public RenamePartitionResponse rename_partition_req( RenamePartitionRequest req) throws InvalidOperationException ,MetaException ,TException { rename_partition(req.getCatName(), req.getDbName(), req.getTableName(), req.getPartVals(), - req.getNewPart(), null, req.getTxnId(), req.getValidWriteIdList()); + req.getNewPart(), null, req.getValidWriteIdList()); return new RenamePartitionResponse(); }; private void rename_partition(String catName, String db_name, String tbl_name, - List<String> part_vals, Partition new_part, EnvironmentContext envContext, long txnId, + List<String> part_vals, Partition new_part, EnvironmentContext envContext, String validWriteIds) throws TException { startTableFunction("alter_partition", catName, db_name, tbl_name); @@ -4883,13 +4882,13 @@ public class HiveMetaStore extends ThriftHiveMetastore { } oldPart = alterHandler.alterPartition(getMS(), wh, catName, db_name, tbl_name, - part_vals, new_part, envContext, this, txnId, validWriteIds); + part_vals, new_part, envContext, this, validWriteIds); // Only fetch the table if we actually have a listener Table table = null; if (!listeners.isEmpty()) { if (table == null) { - table = getMS().getTable(catName, db_name, tbl_name, -1, null); + table = getMS().getTable(catName, db_name, tbl_name, null); } MetaStoreListenerNotifier.notifyEvent(listeners, @@ -4920,14 +4919,13 @@ public class HiveMetaStore extends ThriftHiveMetastore { throws TException { String[] o = parseDbName(db_name, conf); alter_partitions_with_environment_context(o[0], o[1], - tbl_name, new_parts, null, -1, null, -1); + tbl_name, new_parts, null, null, -1); } @Override public AlterPartitionsResponse alter_partitions_req(AlterPartitionsRequest req) throws TException { alter_partitions_with_environment_context(req.getCatName(), req.getDbName(), req.getTableName(), req.getPartitions(), req.getEnvironmentContext(), - req.isSetTxnId() ? req.getTxnId() : -1, req.isSetValidWriteIdList() ? req.getValidWriteIdList() : null, req.isSetWriteId() ? req.getWriteId() : -1); return new AlterPartitionsResponse(); @@ -4941,12 +4939,12 @@ public class HiveMetaStore extends ThriftHiveMetastore { throws TException { String[] o = parseDbName(db_name, conf); alter_partitions_with_environment_context(o[0], o[1], tbl_name, new_parts, environmentContext, - -1, null, -1); + null, -1); } private void alter_partitions_with_environment_context(String catName, String db_name, final String tbl_name, final List<Partition> new_parts, EnvironmentContext environmentContext, - long txnId, String writeIdList, long writeId) + String writeIdList, long writeId) throws TException { if (environmentContext == null) { environmentContext = new EnvironmentContext(); @@ -4975,7 +4973,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { firePreEvent(new PreAlterPartitionEvent(db_name, tbl_name, null, tmpPart, this)); } oldParts = alterHandler.alterPartitions(getMS(), wh, - catName, db_name, tbl_name, new_parts, environmentContext, txnId, writeIdList, writeId, this); + catName, db_name, tbl_name, new_parts, environmentContext, writeIdList, writeId, this); Iterator<Partition> olditr = oldParts.iterator(); // Only fetch the table if we have a listener that needs it. Table table = null; @@ -4989,7 +4987,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } if (table == null) { - table = getMS().getTable(catName, db_name, tbl_name, -1, null); + table = getMS().getTable(catName, db_name, tbl_name, null); } if (!listeners.isEmpty()) { @@ -5028,7 +5026,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { // Do not set an environment context. String[] parsedDbName = parseDbName(dbname, conf); alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable, - null, -1, null); + null, null); } @Override @@ -5042,14 +5040,14 @@ public class HiveMetaStore extends ThriftHiveMetastore { } String[] parsedDbName = parseDbName(dbname, conf); alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], name, newTable, - envContext, -1, null); + envContext, null); } @Override public AlterTableResponse alter_table_req(AlterTableRequest req) throws InvalidOperationException, MetaException, TException { alter_table_core(req.getCatName(), req.getDbName(), req.getTableName(), - req.getTable(), req.getEnvironmentContext(), req.getTxnId(), req.getValidWriteIdList()); + req.getTable(), req.getEnvironmentContext(), req.getValidWriteIdList()); return new AlterTableResponse(); } @@ -5060,11 +5058,11 @@ public class HiveMetaStore extends ThriftHiveMetastore { throws InvalidOperationException, MetaException { String[] parsedDbName = parseDbName(dbname, conf); alter_table_core(parsedDbName[CAT_NAME], parsedDbName[DB_NAME], - name, newTable, envContext, -1, null); + name, newTable, envContext, null); } private void alter_table_core(String catName, String dbname, String name, Table newTable, - EnvironmentContext envContext, long txnId, String validWriteIdList) + EnvironmentContext envContext, String validWriteIdList) throws InvalidOperationException, MetaException { startFunction("alter_table", ": " + TableName.getQualified(catName, dbname, name) + " newtbl=" + newTable.getTableName()); @@ -5101,7 +5099,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { Table oldt = get_table_core(catName, dbname, name); firePreEvent(new PreAlterTableEvent(oldt, newTable, this)); alterHandler.alterTable(getMS(), wh, catName, dbname, name, newTable, - envContext, this, txnId, validWriteIdList); + envContext, this, validWriteIdList); success = true; } catch (NoSuchObjectException e) { // thrown when the table to be altered does not exist @@ -5426,7 +5424,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { private List<String> getPartValsFromName(RawStore ms, String catName, String dbName, String tblName, String partName) throws MetaException, InvalidObjectException { - Table t = ms.getTable(catName, dbName, tblName, -1, null); + Table t = ms.getTable(catName, dbName, tblName, null); if (t == null) { throw new InvalidObjectException(dbName + "." + tblName + " table not found"); @@ -5684,7 +5682,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { statsObj = getMS().getTableColumnStatistics( parsedDbName[CAT_NAME], parsedDbName[DB_NAME], tableName, Lists.newArrayList(colName), - -1, null); + null); if (statsObj != null) { assert statsObj.getStatsObjSize() <= 1; } @@ -5710,7 +5708,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { ColumnStatistics cs = getMS().getTableColumnStatistics( catName, dbName, tblName, lowerCaseColNames, - request.getTxnId(), request.getValidWriteIdList()); + request.getValidWriteIdList()); // Note: stats compliance is not propagated to the client; instead, we just return nothing // if stats are not compliant for now. This won't work for stats merging, but that // is currently only done on metastore size (see set_aggr...). @@ -5779,7 +5777,6 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { List<ColumnStatistics> stats = getMS().getPartitionColumnStatistics( catName, dbName, tblName, lowerCasePartNames, lowerCaseColNames, - request.isSetTxnId() ? request.getTxnId() : -1, request.isSetValidWriteIdList() ? request.getValidWriteIdList() : null); Map<String, List<ColumnStatisticsObj>> map = new HashMap<>(); if (stats != null) { @@ -5803,7 +5800,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { @Override public boolean update_table_column_statistics(ColumnStatistics colStats) throws TException { // Deprecated API, won't work for transactional tables - return updateTableColumnStatsInternal(colStats, -1, null, -1); + return updateTableColumnStatsInternal(colStats, null, -1); } @Override @@ -5819,12 +5816,12 @@ public class HiveMetaStore extends ThriftHiveMetastore { } ColumnStatistics colStats = req.getColStatsIterator().next(); boolean ret = updateTableColumnStatsInternal(colStats, - req.getTxnId(), req.getValidWriteIdList(), req.getWriteId()); + req.getValidWriteIdList(), req.getWriteId()); return new SetPartitionsStatsResponse(ret); } private boolean updateTableColumnStatsInternal(ColumnStatistics colStats, - long txnId, String validWriteIds, long writeId) + String validWriteIds, long writeId) throws NoSuchObjectException, MetaException, InvalidObjectException, InvalidInputException { normalizeColStatsInput(colStats); @@ -5834,7 +5831,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { boolean ret = false; try { - ret = getMS().updateTableColumnStatistics(colStats, txnId, validWriteIds, writeId); + ret = getMS().updateTableColumnStatistics(colStats, validWriteIds, writeId); } finally { endFunction("write_column_statistics", ret != false, null, colStats.getStatsDesc().getTableName()); @@ -5861,7 +5858,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } private boolean updatePartitonColStatsInternal(Table tbl, ColumnStatistics colStats, - long txnId, String validWriteIds, long writeId) + String validWriteIds, long writeId) throws MetaException, InvalidObjectException, NoSuchObjectException, InvalidInputException { normalizeColStatsInput(colStats); @@ -5878,7 +5875,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } List<String> partVals = getPartValsFromName(tbl, csd.getPartName()); return getMS().updatePartitionColumnStatistics( - colStats, partVals, txnId, validWriteIds, writeId); + colStats, partVals, validWriteIds, writeId); } finally { endFunction("write_partition_column_statistics", ret != false, null, tableName); } @@ -5887,7 +5884,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { @Override public boolean update_partition_column_statistics(ColumnStatistics colStats) throws TException { // Deprecated API. - return updatePartitonColStatsInternal(null, colStats, -1, null, -1); + return updatePartitonColStatsInternal(null, colStats, null, -1); } @@ -5904,7 +5901,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } ColumnStatistics colStats = req.getColStatsIterator().next(); boolean ret = updatePartitonColStatsInternal(null, colStats, - req.getTxnId(), req.getValidWriteIdList(), req.getWriteId()); + req.getValidWriteIdList(), req.getWriteId()); return new SetPartitionsStatsResponse(ret); } @@ -7560,8 +7557,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { try { aggrStats = getMS().get_aggr_stats_for(catName, dbName, tblName, - lowerCasePartNames, lowerCaseColNames, request.getTxnId(), - request.getValidWriteIdList()); + lowerCasePartNames, lowerCaseColNames, request.getValidWriteIdList()); return aggrStats; } finally { endFunction("get_aggr_stats_for", aggrStats == null, null, request.getTblName()); @@ -7597,12 +7593,11 @@ public class HiveMetaStore extends ThriftHiveMetastore { return updateTableColumnStatsWithMerge(catName, dbName, tableName, colNames, request); } else { // This is the overwrite case, we do not care about the accuracy. - return updateTableColumnStatsInternal(firstColStats, request.getTxnId(), + return updateTableColumnStatsInternal(firstColStats, request.getValidWriteIdList(), request.getWriteId()); } } else { // partition level column stats merging - List<Partition> partitions = new ArrayList<>(); // note that we may have two or more duplicate partition names. // see autoColumnStats_2.q under TestMiniLlapLocalCliDriver Map<String, ColumnStatistics> newStatsMap = new HashMap<>(); @@ -7621,7 +7616,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { Table t = getTable(catName, dbName, tableName); for (Entry<String, ColumnStatistics> entry : newStatsMap.entrySet()) { // We don't short-circuit on errors here anymore. That can leave acid stats invalid. - ret = updatePartitonColStatsInternal(t, entry.getValue(), request.getTxnId(), + ret = updatePartitonColStatsInternal(t, entry.getValue(), request.getValidWriteIdList(), request.getWriteId()) && ret; } } @@ -7640,7 +7635,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { List<String> partitionNames = new ArrayList<>(); partitionNames.addAll(newStatsMap.keySet()); List<ColumnStatistics> csOlds = ms.getPartitionColumnStatistics(catName, dbName, tableName, - partitionNames, colNames, request.getTxnId(), request.getValidWriteIdList()); + partitionNames, colNames, request.getValidWriteIdList()); if (newStatsMap.values().size() != csOlds.size()) { // some of the partitions miss stats. LOG.debug("Some of the partitions miss stats."); @@ -7678,7 +7673,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { if (!csNew.getStatsObj().isEmpty()) { // We don't short-circuit on errors here anymore. That can leave acid stats invalid. - result = updatePartitonColStatsInternal(t, csNew, request.getTxnId(), + result = updatePartitonColStatsInternal(t, csNew, request.getValidWriteIdList(), request.getWriteId()) && result; } else if (isInvalidTxnStats) { // For now because the stats state is such as it is, we will invalidate everything. @@ -7688,7 +7683,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { StatsSetupConst.clearColumnStatsState(part.getParameters()); StatsSetupConst.setBasicStatsState(part.getParameters(), StatsSetupConst.FALSE); ms.alterPartition(catName, dbName, tableName, part.getValues(), part, - request.getTxnId(), request.getValidWriteIdList()); + request.getValidWriteIdList()); result = false; } else { // TODO: why doesn't the original call for non acid tables invalidate the stats? @@ -7716,7 +7711,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { boolean isCommitted = false, result = false; try { ColumnStatistics csOld = ms.getTableColumnStatistics(catName, dbName, tableName, colNames, - request.getTxnId(), request.getValidWriteIdList()); + request.getValidWriteIdList()); // we first use the valid stats list to prune the stats boolean isInvalidTxnStats = csOld != null && csOld.isSetIsStatsCompliant() && !csOld.isIsStatsCompliant(); @@ -7734,7 +7729,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { } if (!firstColStats.getStatsObj().isEmpty()) { - result = updateTableColumnStatsInternal(firstColStats, request.getTxnId(), + result = updateTableColumnStatsInternal(firstColStats, request.getValidWriteIdList(), request.getWriteId()); } else if (isInvalidTxnStats) { // For now because the stats state is such as it is, we will invalidate everything. @@ -7744,7 +7739,7 @@ public class HiveMetaStore extends ThriftHiveMetastore { t.setWriteId(request.getWriteId()); StatsSetupConst.clearColumnStatsState(t.getParameters()); StatsSetupConst.setBasicStatsState(t.getParameters(), StatsSetupConst.FALSE); - ms.alterTable(catName, dbName, tableName, t, request.getTxnId(), request.getValidWriteIdList()); + ms.alterTable(catName, dbName, tableName, t, request.getValidWriteIdList()); } else { // TODO: why doesn't the original call for non acid tables invalidate the stats? LOG.debug("All the column stats are not accurate to merge."); @@ -7763,13 +7758,13 @@ public class HiveMetaStore extends ThriftHiveMetastore { private Table getTable(String catName, String dbName, String tableName) throws MetaException, InvalidObjectException { - return getTable(catName, dbName, tableName, -1, null); + return getTable(catName, dbName, tableName, null); } private Table getTable(String catName, String dbName, String tableName, - long txnId, String writeIdList) + String writeIdList) throws MetaException, InvalidObjectException { - Table t = getMS().getTable(catName, dbName, tableName, txnId, writeIdList); + Table t = getMS().getTable(catName, dbName, tableName, writeIdList); if (t == null) { throw new InvalidObjectException(TableName.getQualified(catName, dbName, tableName) + " table not found");
