Modified: hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java URL: http://svn.apache.org/viewvc/hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java?rev=1622763&r1=1622762&r2=1622763&view=diff ============================================================================== --- hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java (original) +++ hive/branches/tez/metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/ThriftHiveMetastore.java Fri Sep 5 19:15:44 2014 @@ -94,6 +94,8 @@ public class ThriftHiveMetastore { public int add_partitions(List<Partition> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException; + public int add_partitions_pspec(List<PartitionSpec> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException; + public Partition append_partition(String db_name, String tbl_name, List<String> part_vals) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException; public AddPartitionsResult add_partitions_req(AddPartitionsRequest request) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException; @@ -126,6 +128,8 @@ public class ThriftHiveMetastore { public List<Partition> get_partitions_with_auth(String db_name, String tbl_name, short max_parts, String user_name, List<String> group_names) throws NoSuchObjectException, MetaException, org.apache.thrift.TException; + public List<PartitionSpec> get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws NoSuchObjectException, MetaException, org.apache.thrift.TException; + public List<String> get_partition_names(String db_name, String tbl_name, short max_parts) throws MetaException, org.apache.thrift.TException; public List<Partition> get_partitions_ps(String db_name, String tbl_name, List<String> part_vals, short max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException; @@ -136,6 +140,8 @@ public class ThriftHiveMetastore { public List<Partition> get_partitions_by_filter(String db_name, String tbl_name, String filter, short max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException; + public List<PartitionSpec> get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException; + public PartitionsByExprResult get_partitions_by_expr(PartitionsByExprRequest req) throws MetaException, NoSuchObjectException, org.apache.thrift.TException; public List<Partition> get_partitions_by_names(String db_name, String tbl_name, List<String> names) throws MetaException, NoSuchObjectException, org.apache.thrift.TException; @@ -324,6 +330,8 @@ public class ThriftHiveMetastore { public void add_partitions(List<Partition> new_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_call> resultHandler) throws org.apache.thrift.TException; + public void add_partitions_pspec(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException; + public void append_partition(String db_name, String tbl_name, List<String> part_vals, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.append_partition_call> resultHandler) throws org.apache.thrift.TException; public void add_partitions_req(AddPartitionsRequest request, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.add_partitions_req_call> resultHandler) throws org.apache.thrift.TException; @@ -356,6 +364,8 @@ public class ThriftHiveMetastore { public void get_partitions_with_auth(String db_name, String tbl_name, short max_parts, String user_name, List<String> group_names, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_with_auth_call> resultHandler) throws org.apache.thrift.TException; + public void get_partitions_pspec(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException; + public void get_partition_names(String db_name, String tbl_name, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partition_names_call> resultHandler) throws org.apache.thrift.TException; public void get_partitions_ps(String db_name, String tbl_name, List<String> part_vals, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_ps_call> resultHandler) throws org.apache.thrift.TException; @@ -366,6 +376,8 @@ public class ThriftHiveMetastore { public void get_partitions_by_filter(String db_name, String tbl_name, String filter, short max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_filter_call> resultHandler) throws org.apache.thrift.TException; + public void get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_part_specs_by_filter_call> resultHandler) throws org.apache.thrift.TException; + public void get_partitions_by_expr(PartitionsByExprRequest req, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_expr_call> resultHandler) throws org.apache.thrift.TException; public void get_partitions_by_names(String db_name, String tbl_name, List<String> names, org.apache.thrift.async.AsyncMethodCallback<AsyncClient.get_partitions_by_names_call> resultHandler) throws org.apache.thrift.TException; @@ -1341,6 +1353,38 @@ public class ThriftHiveMetastore { throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_partitions failed: unknown result"); } + public int add_partitions_pspec(List<PartitionSpec> new_parts) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException + { + send_add_partitions_pspec(new_parts); + return recv_add_partitions_pspec(); + } + + public void send_add_partitions_pspec(List<PartitionSpec> new_parts) throws org.apache.thrift.TException + { + add_partitions_pspec_args args = new add_partitions_pspec_args(); + args.setNew_parts(new_parts); + sendBase("add_partitions_pspec", args); + } + + public int recv_add_partitions_pspec() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException + { + add_partitions_pspec_result result = new add_partitions_pspec_result(); + receiveBase(result, "add_partitions_pspec"); + if (result.isSetSuccess()) { + return result.success; + } + if (result.o1 != null) { + throw result.o1; + } + if (result.o2 != null) { + throw result.o2; + } + if (result.o3 != null) { + throw result.o3; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "add_partitions_pspec failed: unknown result"); + } + public Partition append_partition(String db_name, String tbl_name, List<String> part_vals) throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException { send_append_partition(db_name, tbl_name, part_vals); @@ -1868,6 +1912,37 @@ public class ThriftHiveMetastore { throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_with_auth failed: unknown result"); } + public List<PartitionSpec> get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws NoSuchObjectException, MetaException, org.apache.thrift.TException + { + send_get_partitions_pspec(db_name, tbl_name, max_parts); + return recv_get_partitions_pspec(); + } + + public void send_get_partitions_pspec(String db_name, String tbl_name, int max_parts) throws org.apache.thrift.TException + { + get_partitions_pspec_args args = new get_partitions_pspec_args(); + args.setDb_name(db_name); + args.setTbl_name(tbl_name); + args.setMax_parts(max_parts); + sendBase("get_partitions_pspec", args); + } + + public List<PartitionSpec> recv_get_partitions_pspec() throws NoSuchObjectException, MetaException, org.apache.thrift.TException + { + get_partitions_pspec_result result = new get_partitions_pspec_result(); + receiveBase(result, "get_partitions_pspec"); + if (result.isSetSuccess()) { + return result.success; + } + if (result.o1 != null) { + throw result.o1; + } + if (result.o2 != null) { + throw result.o2; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_pspec failed: unknown result"); + } + public List<String> get_partition_names(String db_name, String tbl_name, short max_parts) throws MetaException, org.apache.thrift.TException { send_get_partition_names(db_name, tbl_name, max_parts); @@ -2026,6 +2101,38 @@ public class ThriftHiveMetastore { throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_partitions_by_filter failed: unknown result"); } + public List<PartitionSpec> get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws MetaException, NoSuchObjectException, org.apache.thrift.TException + { + send_get_part_specs_by_filter(db_name, tbl_name, filter, max_parts); + return recv_get_part_specs_by_filter(); + } + + public void send_get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts) throws org.apache.thrift.TException + { + get_part_specs_by_filter_args args = new get_part_specs_by_filter_args(); + args.setDb_name(db_name); + args.setTbl_name(tbl_name); + args.setFilter(filter); + args.setMax_parts(max_parts); + sendBase("get_part_specs_by_filter", args); + } + + public List<PartitionSpec> recv_get_part_specs_by_filter() throws MetaException, NoSuchObjectException, org.apache.thrift.TException + { + get_part_specs_by_filter_result result = new get_part_specs_by_filter_result(); + receiveBase(result, "get_part_specs_by_filter"); + if (result.isSetSuccess()) { + return result.success; + } + if (result.o1 != null) { + throw result.o1; + } + if (result.o2 != null) { + throw result.o2; + } + throw new org.apache.thrift.TApplicationException(org.apache.thrift.TApplicationException.MISSING_RESULT, "get_part_specs_by_filter failed: unknown result"); + } + public PartitionsByExprResult get_partitions_by_expr(PartitionsByExprRequest req) throws MetaException, NoSuchObjectException, org.apache.thrift.TException { send_get_partitions_by_expr(req); @@ -4840,6 +4947,38 @@ public class ThriftHiveMetastore { } } + public void add_partitions_pspec(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<add_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException { + checkReady(); + add_partitions_pspec_call method_call = new add_partitions_pspec_call(new_parts, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + public static class add_partitions_pspec_call extends org.apache.thrift.async.TAsyncMethodCall { + private List<PartitionSpec> new_parts; + public add_partitions_pspec_call(List<PartitionSpec> new_parts, org.apache.thrift.async.AsyncMethodCallback<add_partitions_pspec_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.new_parts = new_parts; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("add_partitions_pspec", org.apache.thrift.protocol.TMessageType.CALL, 0)); + add_partitions_pspec_args args = new add_partitions_pspec_args(); + args.setNew_parts(new_parts); + args.write(prot); + prot.writeMessageEnd(); + } + + public int getResult() throws InvalidObjectException, AlreadyExistsException, MetaException, org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_add_partitions_pspec(); + } + } + public void append_partition(String db_name, String tbl_name, List<String> part_vals, org.apache.thrift.async.AsyncMethodCallback<append_partition_call> resultHandler) throws org.apache.thrift.TException { checkReady(); append_partition_call method_call = new append_partition_call(db_name, tbl_name, part_vals, resultHandler, this, ___protocolFactory, ___transport); @@ -5478,6 +5617,44 @@ public class ThriftHiveMetastore { } } + public void get_partitions_pspec(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partitions_pspec_call> resultHandler) throws org.apache.thrift.TException { + checkReady(); + get_partitions_pspec_call method_call = new get_partitions_pspec_call(db_name, tbl_name, max_parts, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + public static class get_partitions_pspec_call extends org.apache.thrift.async.TAsyncMethodCall { + private String db_name; + private String tbl_name; + private int max_parts; + public get_partitions_pspec_call(String db_name, String tbl_name, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partitions_pspec_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.db_name = db_name; + this.tbl_name = tbl_name; + this.max_parts = max_parts; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("get_partitions_pspec", org.apache.thrift.protocol.TMessageType.CALL, 0)); + get_partitions_pspec_args args = new get_partitions_pspec_args(); + args.setDb_name(db_name); + args.setTbl_name(tbl_name); + args.setMax_parts(max_parts); + args.write(prot); + prot.writeMessageEnd(); + } + + public List<PartitionSpec> getResult() throws NoSuchObjectException, MetaException, org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_get_partitions_pspec(); + } + } + public void get_partition_names(String db_name, String tbl_name, short max_parts, org.apache.thrift.async.AsyncMethodCallback<get_partition_names_call> resultHandler) throws org.apache.thrift.TException { checkReady(); get_partition_names_call method_call = new get_partition_names_call(db_name, tbl_name, max_parts, resultHandler, this, ___protocolFactory, ___transport); @@ -5686,6 +5863,47 @@ public class ThriftHiveMetastore { } } + public void get_part_specs_by_filter(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_part_specs_by_filter_call> resultHandler) throws org.apache.thrift.TException { + checkReady(); + get_part_specs_by_filter_call method_call = new get_part_specs_by_filter_call(db_name, tbl_name, filter, max_parts, resultHandler, this, ___protocolFactory, ___transport); + this.___currentMethod = method_call; + ___manager.call(method_call); + } + + public static class get_part_specs_by_filter_call extends org.apache.thrift.async.TAsyncMethodCall { + private String db_name; + private String tbl_name; + private String filter; + private int max_parts; + public get_part_specs_by_filter_call(String db_name, String tbl_name, String filter, int max_parts, org.apache.thrift.async.AsyncMethodCallback<get_part_specs_by_filter_call> resultHandler, org.apache.thrift.async.TAsyncClient client, org.apache.thrift.protocol.TProtocolFactory protocolFactory, org.apache.thrift.transport.TNonblockingTransport transport) throws org.apache.thrift.TException { + super(client, protocolFactory, transport, resultHandler, false); + this.db_name = db_name; + this.tbl_name = tbl_name; + this.filter = filter; + this.max_parts = max_parts; + } + + public void write_args(org.apache.thrift.protocol.TProtocol prot) throws org.apache.thrift.TException { + prot.writeMessageBegin(new org.apache.thrift.protocol.TMessage("get_part_specs_by_filter", org.apache.thrift.protocol.TMessageType.CALL, 0)); + get_part_specs_by_filter_args args = new get_part_specs_by_filter_args(); + args.setDb_name(db_name); + args.setTbl_name(tbl_name); + args.setFilter(filter); + args.setMax_parts(max_parts); + args.write(prot); + prot.writeMessageEnd(); + } + + public List<PartitionSpec> getResult() throws MetaException, NoSuchObjectException, org.apache.thrift.TException { + if (getState() != org.apache.thrift.async.TAsyncMethodCall.State.RESPONSE_READ) { + throw new IllegalStateException("Method call not finished!"); + } + org.apache.thrift.transport.TMemoryInputTransport memoryTransport = new org.apache.thrift.transport.TMemoryInputTransport(getFrameBuffer().array()); + org.apache.thrift.protocol.TProtocol prot = client.getProtocolFactory().getProtocol(memoryTransport); + return (new Client(prot)).recv_get_part_specs_by_filter(); + } + } + public void get_partitions_by_expr(PartitionsByExprRequest req, org.apache.thrift.async.AsyncMethodCallback<get_partitions_by_expr_call> resultHandler) throws org.apache.thrift.TException { checkReady(); get_partitions_by_expr_call method_call = new get_partitions_by_expr_call(req, resultHandler, this, ___protocolFactory, ___transport); @@ -7952,6 +8170,7 @@ public class ThriftHiveMetastore { processMap.put("add_partition", new add_partition()); processMap.put("add_partition_with_environment_context", new add_partition_with_environment_context()); processMap.put("add_partitions", new add_partitions()); + processMap.put("add_partitions_pspec", new add_partitions_pspec()); processMap.put("append_partition", new append_partition()); processMap.put("add_partitions_req", new add_partitions_req()); processMap.put("append_partition_with_environment_context", new append_partition_with_environment_context()); @@ -7968,11 +8187,13 @@ public class ThriftHiveMetastore { processMap.put("get_partition_by_name", new get_partition_by_name()); processMap.put("get_partitions", new get_partitions()); processMap.put("get_partitions_with_auth", new get_partitions_with_auth()); + processMap.put("get_partitions_pspec", new get_partitions_pspec()); processMap.put("get_partition_names", new get_partition_names()); processMap.put("get_partitions_ps", new get_partitions_ps()); processMap.put("get_partitions_ps_with_auth", new get_partitions_ps_with_auth()); processMap.put("get_partition_names_ps", new get_partition_names_ps()); processMap.put("get_partitions_by_filter", new get_partitions_by_filter()); + processMap.put("get_part_specs_by_filter", new get_part_specs_by_filter()); processMap.put("get_partitions_by_expr", new get_partitions_by_expr()); processMap.put("get_partitions_by_names", new get_partitions_by_names()); processMap.put("alter_partition", new alter_partition()); @@ -8785,6 +9006,35 @@ public class ThriftHiveMetastore { } } + public static class add_partitions_pspec<I extends Iface> extends org.apache.thrift.ProcessFunction<I, add_partitions_pspec_args> { + public add_partitions_pspec() { + super("add_partitions_pspec"); + } + + public add_partitions_pspec_args getEmptyArgsInstance() { + return new add_partitions_pspec_args(); + } + + protected boolean isOneway() { + return false; + } + + public add_partitions_pspec_result getResult(I iface, add_partitions_pspec_args args) throws org.apache.thrift.TException { + add_partitions_pspec_result result = new add_partitions_pspec_result(); + try { + result.success = iface.add_partitions_pspec(args.new_parts); + result.setSuccessIsSet(true); + } catch (InvalidObjectException o1) { + result.o1 = o1; + } catch (AlreadyExistsException o2) { + result.o2 = o2; + } catch (MetaException o3) { + result.o3 = o3; + } + return result; + } + } + public static class append_partition<I extends Iface> extends org.apache.thrift.ProcessFunction<I, append_partition_args> { public append_partition() { super("append_partition"); @@ -9219,6 +9469,32 @@ public class ThriftHiveMetastore { } } + public static class get_partitions_pspec<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partitions_pspec_args> { + public get_partitions_pspec() { + super("get_partitions_pspec"); + } + + public get_partitions_pspec_args getEmptyArgsInstance() { + return new get_partitions_pspec_args(); + } + + protected boolean isOneway() { + return false; + } + + public get_partitions_pspec_result getResult(I iface, get_partitions_pspec_args args) throws org.apache.thrift.TException { + get_partitions_pspec_result result = new get_partitions_pspec_result(); + try { + result.success = iface.get_partitions_pspec(args.db_name, args.tbl_name, args.max_parts); + } catch (NoSuchObjectException o1) { + result.o1 = o1; + } catch (MetaException o2) { + result.o2 = o2; + } + return result; + } + } + public static class get_partition_names<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partition_names_args> { public get_partition_names() { super("get_partition_names"); @@ -9347,6 +9623,32 @@ public class ThriftHiveMetastore { } } + public static class get_part_specs_by_filter<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_part_specs_by_filter_args> { + public get_part_specs_by_filter() { + super("get_part_specs_by_filter"); + } + + public get_part_specs_by_filter_args getEmptyArgsInstance() { + return new get_part_specs_by_filter_args(); + } + + protected boolean isOneway() { + return false; + } + + public get_part_specs_by_filter_result getResult(I iface, get_part_specs_by_filter_args args) throws org.apache.thrift.TException { + get_part_specs_by_filter_result result = new get_part_specs_by_filter_result(); + try { + result.success = iface.get_part_specs_by_filter(args.db_name, args.tbl_name, args.filter, args.max_parts); + } catch (MetaException o1) { + result.o1 = o1; + } catch (NoSuchObjectException o2) { + result.o2 = o2; + } + return result; + } + } + public static class get_partitions_by_expr<I extends Iface> extends org.apache.thrift.ProcessFunction<I, get_partitions_by_expr_args> { public get_partitions_by_expr() { super("get_partitions_by_expr"); @@ -16386,13 +16688,13 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list458 = iprot.readListBegin(); - struct.success = new ArrayList<String>(_list458.size); - for (int _i459 = 0; _i459 < _list458.size; ++_i459) + org.apache.thrift.protocol.TList _list492 = iprot.readListBegin(); + struct.success = new ArrayList<String>(_list492.size); + for (int _i493 = 0; _i493 < _list492.size; ++_i493) { - String _elem460; // required - _elem460 = iprot.readString(); - struct.success.add(_elem460); + String _elem494; // required + _elem494 = iprot.readString(); + struct.success.add(_elem494); } iprot.readListEnd(); } @@ -16427,9 +16729,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter461 : struct.success) + for (String _iter495 : struct.success) { - oprot.writeString(_iter461); + oprot.writeString(_iter495); } oprot.writeListEnd(); } @@ -16468,9 +16770,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter462 : struct.success) + for (String _iter496 : struct.success) { - oprot.writeString(_iter462); + oprot.writeString(_iter496); } } } @@ -16485,13 +16787,13 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list463 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList<String>(_list463.size); - for (int _i464 = 0; _i464 < _list463.size; ++_i464) + org.apache.thrift.protocol.TList _list497 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList<String>(_list497.size); + for (int _i498 = 0; _i498 < _list497.size; ++_i498) { - String _elem465; // required - _elem465 = iprot.readString(); - struct.success.add(_elem465); + String _elem499; // required + _elem499 = iprot.readString(); + struct.success.add(_elem499); } } struct.setSuccessIsSet(true); @@ -17148,13 +17450,13 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list466 = iprot.readListBegin(); - struct.success = new ArrayList<String>(_list466.size); - for (int _i467 = 0; _i467 < _list466.size; ++_i467) + org.apache.thrift.protocol.TList _list500 = iprot.readListBegin(); + struct.success = new ArrayList<String>(_list500.size); + for (int _i501 = 0; _i501 < _list500.size; ++_i501) { - String _elem468; // required - _elem468 = iprot.readString(); - struct.success.add(_elem468); + String _elem502; // required + _elem502 = iprot.readString(); + struct.success.add(_elem502); } iprot.readListEnd(); } @@ -17189,9 +17491,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter469 : struct.success) + for (String _iter503 : struct.success) { - oprot.writeString(_iter469); + oprot.writeString(_iter503); } oprot.writeListEnd(); } @@ -17230,9 +17532,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter470 : struct.success) + for (String _iter504 : struct.success) { - oprot.writeString(_iter470); + oprot.writeString(_iter504); } } } @@ -17247,13 +17549,13 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list471 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList<String>(_list471.size); - for (int _i472 = 0; _i472 < _list471.size; ++_i472) + org.apache.thrift.protocol.TList _list505 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList<String>(_list505.size); + for (int _i506 = 0; _i506 < _list505.size; ++_i506) { - String _elem473; // required - _elem473 = iprot.readString(); - struct.success.add(_elem473); + String _elem507; // required + _elem507 = iprot.readString(); + struct.success.add(_elem507); } } struct.setSuccessIsSet(true); @@ -21860,16 +22162,16 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.MAP) { { - org.apache.thrift.protocol.TMap _map474 = iprot.readMapBegin(); - struct.success = new HashMap<String,Type>(2*_map474.size); - for (int _i475 = 0; _i475 < _map474.size; ++_i475) + org.apache.thrift.protocol.TMap _map508 = iprot.readMapBegin(); + struct.success = new HashMap<String,Type>(2*_map508.size); + for (int _i509 = 0; _i509 < _map508.size; ++_i509) { - String _key476; // required - Type _val477; // required - _key476 = iprot.readString(); - _val477 = new Type(); - _val477.read(iprot); - struct.success.put(_key476, _val477); + String _key510; // required + Type _val511; // required + _key510 = iprot.readString(); + _val511 = new Type(); + _val511.read(iprot); + struct.success.put(_key510, _val511); } iprot.readMapEnd(); } @@ -21904,10 +22206,10 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeMapBegin(new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Map.Entry<String, Type> _iter478 : struct.success.entrySet()) + for (Map.Entry<String, Type> _iter512 : struct.success.entrySet()) { - oprot.writeString(_iter478.getKey()); - _iter478.getValue().write(oprot); + oprot.writeString(_iter512.getKey()); + _iter512.getValue().write(oprot); } oprot.writeMapEnd(); } @@ -21946,10 +22248,10 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Map.Entry<String, Type> _iter479 : struct.success.entrySet()) + for (Map.Entry<String, Type> _iter513 : struct.success.entrySet()) { - oprot.writeString(_iter479.getKey()); - _iter479.getValue().write(oprot); + oprot.writeString(_iter513.getKey()); + _iter513.getValue().write(oprot); } } } @@ -21964,16 +22266,16 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TMap _map480 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new HashMap<String,Type>(2*_map480.size); - for (int _i481 = 0; _i481 < _map480.size; ++_i481) + org.apache.thrift.protocol.TMap _map514 = new org.apache.thrift.protocol.TMap(org.apache.thrift.protocol.TType.STRING, org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new HashMap<String,Type>(2*_map514.size); + for (int _i515 = 0; _i515 < _map514.size; ++_i515) { - String _key482; // required - Type _val483; // required - _key482 = iprot.readString(); - _val483 = new Type(); - _val483.read(iprot); - struct.success.put(_key482, _val483); + String _key516; // required + Type _val517; // required + _key516 = iprot.readString(); + _val517 = new Type(); + _val517.read(iprot); + struct.success.put(_key516, _val517); } } struct.setSuccessIsSet(true); @@ -23008,14 +23310,14 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list484 = iprot.readListBegin(); - struct.success = new ArrayList<FieldSchema>(_list484.size); - for (int _i485 = 0; _i485 < _list484.size; ++_i485) + org.apache.thrift.protocol.TList _list518 = iprot.readListBegin(); + struct.success = new ArrayList<FieldSchema>(_list518.size); + for (int _i519 = 0; _i519 < _list518.size; ++_i519) { - FieldSchema _elem486; // required - _elem486 = new FieldSchema(); - _elem486.read(iprot); - struct.success.add(_elem486); + FieldSchema _elem520; // required + _elem520 = new FieldSchema(); + _elem520.read(iprot); + struct.success.add(_elem520); } iprot.readListEnd(); } @@ -23068,9 +23370,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter487 : struct.success) + for (FieldSchema _iter521 : struct.success) { - _iter487.write(oprot); + _iter521.write(oprot); } oprot.writeListEnd(); } @@ -23125,9 +23427,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter488 : struct.success) + for (FieldSchema _iter522 : struct.success) { - _iter488.write(oprot); + _iter522.write(oprot); } } } @@ -23148,14 +23450,14 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list489 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList<FieldSchema>(_list489.size); - for (int _i490 = 0; _i490 < _list489.size; ++_i490) + org.apache.thrift.protocol.TList _list523 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList<FieldSchema>(_list523.size); + for (int _i524 = 0; _i524 < _list523.size; ++_i524) { - FieldSchema _elem491; // required - _elem491 = new FieldSchema(); - _elem491.read(iprot); - struct.success.add(_elem491); + FieldSchema _elem525; // required + _elem525 = new FieldSchema(); + _elem525.read(iprot); + struct.success.add(_elem525); } } struct.setSuccessIsSet(true); @@ -24200,14 +24502,14 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list492 = iprot.readListBegin(); - struct.success = new ArrayList<FieldSchema>(_list492.size); - for (int _i493 = 0; _i493 < _list492.size; ++_i493) + org.apache.thrift.protocol.TList _list526 = iprot.readListBegin(); + struct.success = new ArrayList<FieldSchema>(_list526.size); + for (int _i527 = 0; _i527 < _list526.size; ++_i527) { - FieldSchema _elem494; // required - _elem494 = new FieldSchema(); - _elem494.read(iprot); - struct.success.add(_elem494); + FieldSchema _elem528; // required + _elem528 = new FieldSchema(); + _elem528.read(iprot); + struct.success.add(_elem528); } iprot.readListEnd(); } @@ -24260,9 +24562,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (FieldSchema _iter495 : struct.success) + for (FieldSchema _iter529 : struct.success) { - _iter495.write(oprot); + _iter529.write(oprot); } oprot.writeListEnd(); } @@ -24317,9 +24619,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (FieldSchema _iter496 : struct.success) + for (FieldSchema _iter530 : struct.success) { - _iter496.write(oprot); + _iter530.write(oprot); } } } @@ -24340,14 +24642,14 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list497 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList<FieldSchema>(_list497.size); - for (int _i498 = 0; _i498 < _list497.size; ++_i498) + org.apache.thrift.protocol.TList _list531 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList<FieldSchema>(_list531.size); + for (int _i532 = 0; _i532 < _list531.size; ++_i532) { - FieldSchema _elem499; // required - _elem499 = new FieldSchema(); - _elem499.read(iprot); - struct.success.add(_elem499); + FieldSchema _elem533; // required + _elem533 = new FieldSchema(); + _elem533.read(iprot); + struct.success.add(_elem533); } } struct.setSuccessIsSet(true); @@ -29590,13 +29892,13 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list500 = iprot.readListBegin(); - struct.success = new ArrayList<String>(_list500.size); - for (int _i501 = 0; _i501 < _list500.size; ++_i501) + org.apache.thrift.protocol.TList _list534 = iprot.readListBegin(); + struct.success = new ArrayList<String>(_list534.size); + for (int _i535 = 0; _i535 < _list534.size; ++_i535) { - String _elem502; // required - _elem502 = iprot.readString(); - struct.success.add(_elem502); + String _elem536; // required + _elem536 = iprot.readString(); + struct.success.add(_elem536); } iprot.readListEnd(); } @@ -29631,9 +29933,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter503 : struct.success) + for (String _iter537 : struct.success) { - oprot.writeString(_iter503); + oprot.writeString(_iter537); } oprot.writeListEnd(); } @@ -29672,9 +29974,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter504 : struct.success) + for (String _iter538 : struct.success) { - oprot.writeString(_iter504); + oprot.writeString(_iter538); } } } @@ -29689,13 +29991,13 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list505 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList<String>(_list505.size); - for (int _i506 = 0; _i506 < _list505.size; ++_i506) + org.apache.thrift.protocol.TList _list539 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList<String>(_list539.size); + for (int _i540 = 0; _i540 < _list539.size; ++_i540) { - String _elem507; // required - _elem507 = iprot.readString(); - struct.success.add(_elem507); + String _elem541; // required + _elem541 = iprot.readString(); + struct.success.add(_elem541); } } struct.setSuccessIsSet(true); @@ -30464,13 +30766,13 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list508 = iprot.readListBegin(); - struct.success = new ArrayList<String>(_list508.size); - for (int _i509 = 0; _i509 < _list508.size; ++_i509) + org.apache.thrift.protocol.TList _list542 = iprot.readListBegin(); + struct.success = new ArrayList<String>(_list542.size); + for (int _i543 = 0; _i543 < _list542.size; ++_i543) { - String _elem510; // required - _elem510 = iprot.readString(); - struct.success.add(_elem510); + String _elem544; // required + _elem544 = iprot.readString(); + struct.success.add(_elem544); } iprot.readListEnd(); } @@ -30505,9 +30807,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter511 : struct.success) + for (String _iter545 : struct.success) { - oprot.writeString(_iter511); + oprot.writeString(_iter545); } oprot.writeListEnd(); } @@ -30546,9 +30848,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter512 : struct.success) + for (String _iter546 : struct.success) { - oprot.writeString(_iter512); + oprot.writeString(_iter546); } } } @@ -30563,13 +30865,13 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(2); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list513 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList<String>(_list513.size); - for (int _i514 = 0; _i514 < _list513.size; ++_i514) + org.apache.thrift.protocol.TList _list547 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList<String>(_list547.size); + for (int _i548 = 0; _i548 < _list547.size; ++_i548) { - String _elem515; // required - _elem515 = iprot.readString(); - struct.success.add(_elem515); + String _elem549; // required + _elem549 = iprot.readString(); + struct.success.add(_elem549); } } struct.setSuccessIsSet(true); @@ -32025,13 +32327,13 @@ public class ThriftHiveMetastore { case 2: // TBL_NAMES if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list516 = iprot.readListBegin(); - struct.tbl_names = new ArrayList<String>(_list516.size); - for (int _i517 = 0; _i517 < _list516.size; ++_i517) + org.apache.thrift.protocol.TList _list550 = iprot.readListBegin(); + struct.tbl_names = new ArrayList<String>(_list550.size); + for (int _i551 = 0; _i551 < _list550.size; ++_i551) { - String _elem518; // required - _elem518 = iprot.readString(); - struct.tbl_names.add(_elem518); + String _elem552; // required + _elem552 = iprot.readString(); + struct.tbl_names.add(_elem552); } iprot.readListEnd(); } @@ -32062,9 +32364,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(TBL_NAMES_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.tbl_names.size())); - for (String _iter519 : struct.tbl_names) + for (String _iter553 : struct.tbl_names) { - oprot.writeString(_iter519); + oprot.writeString(_iter553); } oprot.writeListEnd(); } @@ -32101,9 +32403,9 @@ public class ThriftHiveMetastore { if (struct.isSetTbl_names()) { { oprot.writeI32(struct.tbl_names.size()); - for (String _iter520 : struct.tbl_names) + for (String _iter554 : struct.tbl_names) { - oprot.writeString(_iter520); + oprot.writeString(_iter554); } } } @@ -32119,13 +32421,13 @@ public class ThriftHiveMetastore { } if (incoming.get(1)) { { - org.apache.thrift.protocol.TList _list521 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.tbl_names = new ArrayList<String>(_list521.size); - for (int _i522 = 0; _i522 < _list521.size; ++_i522) + org.apache.thrift.protocol.TList _list555 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.tbl_names = new ArrayList<String>(_list555.size); + for (int _i556 = 0; _i556 < _list555.size; ++_i556) { - String _elem523; // required - _elem523 = iprot.readString(); - struct.tbl_names.add(_elem523); + String _elem557; // required + _elem557 = iprot.readString(); + struct.tbl_names.add(_elem557); } } struct.setTbl_namesIsSet(true); @@ -32693,14 +32995,14 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list524 = iprot.readListBegin(); - struct.success = new ArrayList<Table>(_list524.size); - for (int _i525 = 0; _i525 < _list524.size; ++_i525) + org.apache.thrift.protocol.TList _list558 = iprot.readListBegin(); + struct.success = new ArrayList<Table>(_list558.size); + for (int _i559 = 0; _i559 < _list558.size; ++_i559) { - Table _elem526; // required - _elem526 = new Table(); - _elem526.read(iprot); - struct.success.add(_elem526); + Table _elem560; // required + _elem560 = new Table(); + _elem560.read(iprot); + struct.success.add(_elem560); } iprot.readListEnd(); } @@ -32753,9 +33055,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.success.size())); - for (Table _iter527 : struct.success) + for (Table _iter561 : struct.success) { - _iter527.write(oprot); + _iter561.write(oprot); } oprot.writeListEnd(); } @@ -32810,9 +33112,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (Table _iter528 : struct.success) + for (Table _iter562 : struct.success) { - _iter528.write(oprot); + _iter562.write(oprot); } } } @@ -32833,14 +33135,14 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list529 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.success = new ArrayList<Table>(_list529.size); - for (int _i530 = 0; _i530 < _list529.size; ++_i530) + org.apache.thrift.protocol.TList _list563 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.success = new ArrayList<Table>(_list563.size); + for (int _i564 = 0; _i564 < _list563.size; ++_i564) { - Table _elem531; // required - _elem531 = new Table(); - _elem531.read(iprot); - struct.success.add(_elem531); + Table _elem565; // required + _elem565 = new Table(); + _elem565.read(iprot); + struct.success.add(_elem565); } } struct.setSuccessIsSet(true); @@ -33989,13 +34291,13 @@ public class ThriftHiveMetastore { case 0: // SUCCESS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list532 = iprot.readListBegin(); - struct.success = new ArrayList<String>(_list532.size); - for (int _i533 = 0; _i533 < _list532.size; ++_i533) + org.apache.thrift.protocol.TList _list566 = iprot.readListBegin(); + struct.success = new ArrayList<String>(_list566.size); + for (int _i567 = 0; _i567 < _list566.size; ++_i567) { - String _elem534; // required - _elem534 = iprot.readString(); - struct.success.add(_elem534); + String _elem568; // required + _elem568 = iprot.readString(); + struct.success.add(_elem568); } iprot.readListEnd(); } @@ -34048,9 +34350,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(SUCCESS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.success.size())); - for (String _iter535 : struct.success) + for (String _iter569 : struct.success) { - oprot.writeString(_iter535); + oprot.writeString(_iter569); } oprot.writeListEnd(); } @@ -34105,9 +34407,9 @@ public class ThriftHiveMetastore { if (struct.isSetSuccess()) { { oprot.writeI32(struct.success.size()); - for (String _iter536 : struct.success) + for (String _iter570 : struct.success) { - oprot.writeString(_iter536); + oprot.writeString(_iter570); } } } @@ -34128,13 +34430,13 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(4); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list537 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.success = new ArrayList<String>(_list537.size); - for (int _i538 = 0; _i538 < _list537.size; ++_i538) + org.apache.thrift.protocol.TList _list571 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); + struct.success = new ArrayList<String>(_list571.size); + for (int _i572 = 0; _i572 < _list571.size; ++_i572) { - String _elem539; // required - _elem539 = iprot.readString(); - struct.success.add(_elem539); + String _elem573; // required + _elem573 = iprot.readString(); + struct.success.add(_elem573); } } struct.setSuccessIsSet(true); @@ -38854,14 +39156,14 @@ public class ThriftHiveMetastore { case 1: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list540 = iprot.readListBegin(); - struct.new_parts = new ArrayList<Partition>(_list540.size); - for (int _i541 = 0; _i541 < _list540.size; ++_i541) + org.apache.thrift.protocol.TList _list574 = iprot.readListBegin(); + struct.new_parts = new ArrayList<Partition>(_list574.size); + for (int _i575 = 0; _i575 < _list574.size; ++_i575) { - Partition _elem542; // required - _elem542 = new Partition(); - _elem542.read(iprot); - struct.new_parts.add(_elem542); + Partition _elem576; // required + _elem576 = new Partition(); + _elem576.read(iprot); + struct.new_parts.add(_elem576); } iprot.readListEnd(); } @@ -38887,9 +39189,9 @@ public class ThriftHiveMetastore { oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); - for (Partition _iter543 : struct.new_parts) + for (Partition _iter577 : struct.new_parts) { - _iter543.write(oprot); + _iter577.write(oprot); } oprot.writeListEnd(); } @@ -38920,9 +39222,9 @@ public class ThriftHiveMetastore { if (struct.isSetNew_parts()) { { oprot.writeI32(struct.new_parts.size()); - for (Partition _iter544 : struct.new_parts) + for (Partition _iter578 : struct.new_parts) { - _iter544.write(oprot); + _iter578.write(oprot); } } } @@ -38934,14 +39236,14 @@ public class ThriftHiveMetastore { BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { { - org.apache.thrift.protocol.TList _list545 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); - struct.new_parts = new ArrayList<Partition>(_list545.size); - for (int _i546 = 0; _i546 < _list545.size; ++_i546) + org.apache.thrift.protocol.TList _list579 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList<Partition>(_list579.size); + for (int _i580 = 0; _i580 < _list579.size; ++_i580) { - Partition _elem547; // required - _elem547 = new Partition(); - _elem547.read(iprot); - struct.new_parts.add(_elem547); + Partition _elem581; // required + _elem581 = new Partition(); + _elem581.read(iprot); + struct.new_parts.add(_elem581); } } struct.setNew_partsIsSet(true); @@ -39627,28 +39929,22 @@ public class ThriftHiveMetastore { } - public static class append_partition_args implements org.apache.thrift.TBase<append_partition_args, append_partition_args._Fields>, java.io.Serializable, Cloneable { - private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_args"); + public static class add_partitions_pspec_args implements org.apache.thrift.TBase<add_partitions_pspec_args, add_partitions_pspec_args._Fields>, java.io.Serializable, Cloneable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_pspec_args"); - private static final org.apache.thrift.protocol.TField DB_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("db_name", org.apache.thrift.protocol.TType.STRING, (short)1); - private static final org.apache.thrift.protocol.TField TBL_NAME_FIELD_DESC = new org.apache.thrift.protocol.TField("tbl_name", org.apache.thrift.protocol.TType.STRING, (short)2); - private static final org.apache.thrift.protocol.TField PART_VALS_FIELD_DESC = new org.apache.thrift.protocol.TField("part_vals", org.apache.thrift.protocol.TType.LIST, (short)3); + private static final org.apache.thrift.protocol.TField NEW_PARTS_FIELD_DESC = new org.apache.thrift.protocol.TField("new_parts", org.apache.thrift.protocol.TType.LIST, (short)1); private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>(); static { - schemes.put(StandardScheme.class, new append_partition_argsStandardSchemeFactory()); - schemes.put(TupleScheme.class, new append_partition_argsTupleSchemeFactory()); + schemes.put(StandardScheme.class, new add_partitions_pspec_argsStandardSchemeFactory()); + schemes.put(TupleScheme.class, new add_partitions_pspec_argsTupleSchemeFactory()); } - private String db_name; // required - private String tbl_name; // required - private List<String> part_vals; // required + private List<PartitionSpec> new_parts; // required /** The set of fields this struct contains, along with convenience methods for finding and manipulating them. */ public enum _Fields implements org.apache.thrift.TFieldIdEnum { - DB_NAME((short)1, "db_name"), - TBL_NAME((short)2, "tbl_name"), - PART_VALS((short)3, "part_vals"); + NEW_PARTS((short)1, "new_parts"); private static final Map<String, _Fields> byName = new HashMap<String, _Fields>(); @@ -39663,12 +39959,8 @@ public class ThriftHiveMetastore { */ public static _Fields findByThriftId(int fieldId) { switch(fieldId) { - case 1: // DB_NAME - return DB_NAME; - case 2: // TBL_NAME - return TBL_NAME; - case 3: // PART_VALS - return PART_VALS; + case 1: // NEW_PARTS + return NEW_PARTS; default: return null; } @@ -39712,168 +40004,90 @@ public class ThriftHiveMetastore { public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); - tmpMap.put(_Fields.DB_NAME, new org.apache.thrift.meta_data.FieldMetaData("db_name", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); - tmpMap.put(_Fields.TBL_NAME, new org.apache.thrift.meta_data.FieldMetaData("tbl_name", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING))); - tmpMap.put(_Fields.PART_VALS, new org.apache.thrift.meta_data.FieldMetaData("part_vals", org.apache.thrift.TFieldRequirementType.DEFAULT, + tmpMap.put(_Fields.NEW_PARTS, new org.apache.thrift.meta_data.FieldMetaData("new_parts", org.apache.thrift.TFieldRequirementType.DEFAULT, new org.apache.thrift.meta_data.ListMetaData(org.apache.thrift.protocol.TType.LIST, - new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.STRING)))); + new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, PartitionSpec.class)))); metaDataMap = Collections.unmodifiableMap(tmpMap); - org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(append_partition_args.class, metaDataMap); + org.apache.thrift.meta_data.FieldMetaData.addStructMetaDataMap(add_partitions_pspec_args.class, metaDataMap); } - public append_partition_args() { + public add_partitions_pspec_args() { } - public append_partition_args( - String db_name, - String tbl_name, - List<String> part_vals) + public add_partitions_pspec_args( + List<PartitionSpec> new_parts) { this(); - this.db_name = db_name; - this.tbl_name = tbl_name; - this.part_vals = part_vals; + this.new_parts = new_parts; } /** * Performs a deep copy on <i>other</i>. */ - public append_partition_args(append_partition_args other) { - if (other.isSetDb_name()) { - this.db_name = other.db_name; - } - if (other.isSetTbl_name()) { - this.tbl_name = other.tbl_name; - } - if (other.isSetPart_vals()) { - List<String> __this__part_vals = new ArrayList<String>(); - for (String other_element : other.part_vals) { - __this__part_vals.add(other_element); + public add_partitions_pspec_args(add_partitions_pspec_args other) { + if (other.isSetNew_parts()) { + List<PartitionSpec> __this__new_parts = new ArrayList<PartitionSpec>(); + for (PartitionSpec other_element : other.new_parts) { + __this__new_parts.add(new PartitionSpec(other_element)); } - this.part_vals = __this__part_vals; + this.new_parts = __this__new_parts; } } - public append_partition_args deepCopy() { - return new append_partition_args(this); + public add_partitions_pspec_args deepCopy() { + return new add_partitions_pspec_args(this); } @Override public void clear() { - this.db_name = null; - this.tbl_name = null; - this.part_vals = null; - } - - public String getDb_name() { - return this.db_name; - } - - public void setDb_name(String db_name) { - this.db_name = db_name; - } - - public void unsetDb_name() { - this.db_name = null; - } - - /** Returns true if field db_name is set (has been assigned a value) and false otherwise */ - public boolean isSetDb_name() { - return this.db_name != null; - } - - public void setDb_nameIsSet(boolean value) { - if (!value) { - this.db_name = null; - } - } - - public String getTbl_name() { - return this.tbl_name; - } - - public void setTbl_name(String tbl_name) { - this.tbl_name = tbl_name; - } - - public void unsetTbl_name() { - this.tbl_name = null; - } - - /** Returns true if field tbl_name is set (has been assigned a value) and false otherwise */ - public boolean isSetTbl_name() { - return this.tbl_name != null; - } - - public void setTbl_nameIsSet(boolean value) { - if (!value) { - this.tbl_name = null; - } + this.new_parts = null; } - public int getPart_valsSize() { - return (this.part_vals == null) ? 0 : this.part_vals.size(); + public int getNew_partsSize() { + return (this.new_parts == null) ? 0 : this.new_parts.size(); } - public java.util.Iterator<String> getPart_valsIterator() { - return (this.part_vals == null) ? null : this.part_vals.iterator(); + public java.util.Iterator<PartitionSpec> getNew_partsIterator() { + return (this.new_parts == null) ? null : this.new_parts.iterator(); } - public void addToPart_vals(String elem) { - if (this.part_vals == null) { - this.part_vals = new ArrayList<String>(); + public void addToNew_parts(PartitionSpec elem) { + if (this.new_parts == null) { + this.new_parts = new ArrayList<PartitionSpec>(); } - this.part_vals.add(elem); + this.new_parts.add(elem); } - public List<String> getPart_vals() { - return this.part_vals; + public List<PartitionSpec> getNew_parts() { + return this.new_parts; } - public void setPart_vals(List<String> part_vals) { - this.part_vals = part_vals; + public void setNew_parts(List<PartitionSpec> new_parts) { + this.new_parts = new_parts; } - public void unsetPart_vals() { - this.part_vals = null; + public void unsetNew_parts() { + this.new_parts = null; } - /** Returns true if field part_vals is set (has been assigned a value) and false otherwise */ - public boolean isSetPart_vals() { - return this.part_vals != null; + /** Returns true if field new_parts is set (has been assigned a value) and false otherwise */ + public boolean isSetNew_parts() { + return this.new_parts != null; } - public void setPart_valsIsSet(boolean value) { + public void setNew_partsIsSet(boolean value) { if (!value) { - this.part_vals = null; + this.new_parts = null; } } public void setFieldValue(_Fields field, Object value) { switch (field) { - case DB_NAME: - if (value == null) { - unsetDb_name(); - } else { - setDb_name((String)value); - } - break; - - case TBL_NAME: - if (value == null) { - unsetTbl_name(); - } else { - setTbl_name((String)value); - } - break; - - case PART_VALS: + case NEW_PARTS: if (value == null) { - unsetPart_vals(); + unsetNew_parts(); } else { - setPart_vals((List<String>)value); + setNew_parts((List<PartitionSpec>)value); } break; @@ -39882,14 +40096,8 @@ public class ThriftHiveMetastore { public Object getFieldValue(_Fields field) { switch (field) { - case DB_NAME: - return getDb_name(); - - case TBL_NAME: - return getTbl_name(); - - case PART_VALS: - return getPart_vals(); + case NEW_PARTS: + return getNew_parts(); } throw new IllegalStateException(); @@ -39902,12 +40110,8 @@ public class ThriftHiveMetastore { } switch (field) { - case DB_NAME: - return isSetDb_name(); - case TBL_NAME: - return isSetTbl_name(); - case PART_VALS: - return isSetPart_vals(); + case NEW_PARTS: + return isSetNew_parts(); } throw new IllegalStateException(); } @@ -39916,39 +40120,21 @@ public class ThriftHiveMetastore { public boolean equals(Object that) { if (that == null) return false; - if (that instanceof append_partition_args) - return this.equals((append_partition_args)that); + if (that instanceof add_partitions_pspec_args) + return this.equals((add_partitions_pspec_args)that); return false; } - public boolean equals(append_partition_args that) { + public boolean equals(add_partitions_pspec_args that) { if (that == null) return false; - boolean this_present_db_name = true && this.isSetDb_name(); - boolean that_present_db_name = true && that.isSetDb_name(); - if (this_present_db_name || that_present_db_name) { - if (!(this_present_db_name && that_present_db_name)) - return false; - if (!this.db_name.equals(that.db_name)) - return false; - } - - boolean this_present_tbl_name = true && this.isSetTbl_name(); - boolean that_present_tbl_name = true && that.isSetTbl_name(); - if (this_present_tbl_name || that_present_tbl_name) { - if (!(this_present_tbl_name && that_present_tbl_name)) - return false; - if (!this.tbl_name.equals(that.tbl_name)) - return false; - } - - boolean this_present_part_vals = true && this.isSetPart_vals(); - boolean that_present_part_vals = true && that.isSetPart_vals(); - if (this_present_part_vals || that_present_part_vals) { - if (!(this_present_part_vals && that_present_part_vals)) + boolean this_present_new_parts = true && this.isSetNew_parts(); + boolean that_present_new_parts = true && that.isSetNew_parts(); + if (this_present_new_parts || that_present_new_parts) { + if (!(this_present_new_parts && that_present_new_parts)) return false; - if (!this.part_vals.equals(that.part_vals)) + if (!this.new_parts.equals(that.new_parts)) return false; } @@ -39959,58 +40145,28 @@ public class ThriftHiveMetastore { public int hashCode() { HashCodeBuilder builder = new HashCodeBuilder(); - boolean present_db_name = true && (isSetDb_name()); - builder.append(present_db_name); - if (present_db_name) - builder.append(db_name); - - boolean present_tbl_name = true && (isSetTbl_name()); - builder.append(present_tbl_name); - if (present_tbl_name) - builder.append(tbl_name); - - boolean present_part_vals = true && (isSetPart_vals()); - builder.append(present_part_vals); - if (present_part_vals) - builder.append(part_vals); + boolean present_new_parts = true && (isSetNew_parts()); + builder.append(present_new_parts); + if (present_new_parts) + builder.append(new_parts); return builder.toHashCode(); } - public int compareTo(append_partition_args other) { + public int compareTo(add_partitions_pspec_args other) { if (!getClass().equals(other.getClass())) { return getClass().getName().compareTo(other.getClass().getName()); } int lastComparison = 0; - append_partition_args typedOther = (append_partition_args)other; + add_partitions_pspec_args typedOther = (add_partitions_pspec_args)other; - lastComparison = Boolean.valueOf(isSetDb_name()).compareTo(typedOther.isSetDb_name()); - if (lastComparison != 0) { - return lastComparison; - } - if (isSetDb_name()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.db_name, typedOther.db_name); - if (lastComparison != 0) { - return lastComparison; - } - } - lastComparison = Boolean.valueOf(isSetTbl_name()).compareTo(typedOther.isSetTbl_name()); - if (lastComparison != 0) { - return lastComparison; - } - if (isSetTbl_name()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.tbl_name, typedOther.tbl_name); - if (lastComparison != 0) { - return lastComparison; - } - } - lastComparison = Boolean.valueOf(isSetPart_vals()).compareTo(typedOther.isSetPart_vals()); + lastComparison = Boolean.valueOf(isSetNew_parts()).compareTo(typedOther.isSetNew_parts()); if (lastComparison != 0) { return lastComparison; } - if (isSetPart_vals()) { - lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.part_vals, typedOther.part_vals); + if (isSetNew_parts()) { + lastComparison = org.apache.thrift.TBaseHelper.compareTo(this.new_parts, typedOther.new_parts); if (lastComparison != 0) { return lastComparison; } @@ -40032,30 +40188,14 @@ public class ThriftHiveMetastore { @Override public String toString() { - StringBuilder sb = new StringBuilder("append_partition_args("); + StringBuilder sb = new StringBuilder("add_partitions_pspec_args("); boolean first = true; - sb.append("db_name:"); - if (this.db_name == null) { - sb.append("null"); - } else { - sb.append(this.db_name); - } - first = false; - if (!first) sb.append(", "); - sb.append("tbl_name:"); - if (this.tbl_name == null) { - sb.append("null"); - } else { - sb.append(this.tbl_name); - } - first = false; - if (!first) sb.append(", "); - sb.append("part_vals:"); - if (this.part_vals == null) { + sb.append("new_parts:"); + if (this.new_parts == null) { sb.append("null"); } else { - sb.append(this.part_vals); + sb.append(this.new_parts); } first = false; sb.append(")"); @@ -40083,15 +40223,15 @@ public class ThriftHiveMetastore { } } - private static class append_partition_argsStandardSchemeFactory implements SchemeFactory { - public append_partition_argsStandardScheme getScheme() { - return new append_partition_argsStandardScheme(); + private static class add_partitions_pspec_argsStandardSchemeFactory implements SchemeFactory { + public add_partitions_pspec_argsStandardScheme getScheme() { + return new add_partitions_pspec_argsStandardScheme(); } } - private static class append_partition_argsStandardScheme extends StandardScheme<append_partition_args> { + private static class add_partitions_pspec_argsStandardScheme extends StandardScheme<add_partitions_pspec_args> { - public void read(org.apache.thrift.protocol.TProtocol iprot, append_partition_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol iprot, add_partitions_pspec_args struct) throws org.apache.thrift.TException { org.apache.thrift.protocol.TField schemeField; iprot.readStructBegin(); while (true) @@ -40101,36 +40241,21 @@ public class ThriftHiveMetastore { break; } switch (schemeField.id) { - case 1: // DB_NAME - if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { - struct.db_name = iprot.readString(); - struct.setDb_nameIsSet(true); - } else { - org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); - } - break; - case 2: // TBL_NAME - if (schemeField.type == org.apache.thrift.protocol.TType.STRING) { - struct.tbl_name = iprot.readString(); - struct.setTbl_nameIsSet(true); - } else { - org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); - } - break; - case 3: // PART_VALS + case 1: // NEW_PARTS if (schemeField.type == org.apache.thrift.protocol.TType.LIST) { { - org.apache.thrift.protocol.TList _list548 = iprot.readListBegin(); - struct.part_vals = new ArrayList<String>(_list548.size); - for (int _i549 = 0; _i549 < _list548.size; ++_i549) + org.apache.thrift.protocol.TList _list582 = iprot.readListBegin(); + struct.new_parts = new ArrayList<PartitionSpec>(_list582.size); + for (int _i583 = 0; _i583 < _list582.size; ++_i583) { - String _elem550; // required - _elem550 = iprot.readString(); - struct.part_vals.add(_elem550); + PartitionSpec _elem584; // required + _elem584 = new PartitionSpec(); + _elem584.read(iprot); + struct.new_parts.add(_elem584); } iprot.readListEnd(); } - struct.setPart_valsIsSet(true); + struct.setNew_partsIsSet(true); } else { org.apache.thrift.protocol.TProtocolUtil.skip(iprot, schemeField.type); } @@ -40144,27 +40269,17 @@ public class ThriftHiveMetastore { struct.validate(); } - public void write(org.apache.thrift.protocol.TProtocol oprot, append_partition_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol oprot, add_partitions_pspec_args struct) throws org.apache.thrift.TException { struct.validate(); oprot.writeStructBegin(STRUCT_DESC); - if (struct.db_name != null) { - oprot.writeFieldBegin(DB_NAME_FIELD_DESC); - oprot.writeString(struct.db_name); - oprot.writeFieldEnd(); - } - if (struct.tbl_name != null) { - oprot.writeFieldBegin(TBL_NAME_FIELD_DESC); - oprot.writeString(struct.tbl_name); - oprot.writeFieldEnd(); - } - if (struct.part_vals != null) { - oprot.writeFieldBegin(PART_VALS_FIELD_DESC); + if (struct.new_parts != null) { + oprot.writeFieldBegin(NEW_PARTS_FIELD_DESC); { - oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, struct.part_vals.size())); - for (String _iter551 : struct.part_vals) + oprot.writeListBegin(new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, struct.new_parts.size())); + for (PartitionSpec _iter585 : struct.new_parts) { - oprot.writeString(_iter551); + _iter585.write(oprot); } oprot.writeListEnd(); } @@ -40176,90 +40291,71 @@ public class ThriftHiveMetastore { } - private static class append_partition_argsTupleSchemeFactory implements SchemeFactory { - public append_partition_argsTupleScheme getScheme() { - return new append_partition_argsTupleScheme(); + private static class add_partitions_pspec_argsTupleSchemeFactory implements SchemeFactory { + public add_partitions_pspec_argsTupleScheme getScheme() { + return new add_partitions_pspec_argsTupleScheme(); } } - private static class append_partition_argsTupleScheme extends TupleScheme<append_partition_args> { + private static class add_partitions_pspec_argsTupleScheme extends TupleScheme<add_partitions_pspec_args> { @Override - public void write(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException { + public void write(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspec_args struct) throws org.apache.thrift.TException { TTupleProtocol oprot = (TTupleProtocol) prot; BitSet optionals = new BitSet(); - if (struct.isSetDb_name()) { + if (struct.isSetNew_parts()) { optionals.set(0); } - if (struct.isSetTbl_name()) { - optionals.set(1); - } - if (struct.isSetPart_vals()) { - optionals.set(2); - } - oprot.writeBitSet(optionals, 3); - if (struct.isSetDb_name()) { - oprot.writeString(struct.db_name); - } - if (struct.isSetTbl_name()) { - oprot.writeString(struct.tbl_name); - } - if (struct.isSetPart_vals()) { + oprot.writeBitSet(optionals, 1); + if (struct.isSetNew_parts()) { { - oprot.writeI32(struct.part_vals.size()); - for (String _iter552 : struct.part_vals) + oprot.writeI32(struct.new_parts.size()); + for (PartitionSpec _iter586 : struct.new_parts) { - oprot.writeString(_iter552); + _iter586.write(oprot); } } } } @Override - public void read(org.apache.thrift.protocol.TProtocol prot, append_partition_args struct) throws org.apache.thrift.TException { + public void read(org.apache.thrift.protocol.TProtocol prot, add_partitions_pspec_args struct) throws org.apache.thrift.TException { TTupleProtocol iprot = (TTupleProtocol) prot; - BitSet incoming = iprot.readBitSet(3); + BitSet incoming = iprot.readBitSet(1); if (incoming.get(0)) { - struct.db_name = iprot.readString(); - struct.setDb_nameIsSet(true); - } - if (incoming.get(1)) { - struct.tbl_name = iprot.readString(); - struct.setTbl_nameIsSet(true); - } - if (incoming.get(2)) { { - org.apache.thrift.protocol.TList _list553 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRING, iprot.readI32()); - struct.part_vals = new ArrayList<String>(_list553.size); - for (int _i554 = 0; _i554 < _list553.size; ++_i554) + org.apache.thrift.protocol.TList _list587 = new org.apache.thrift.protocol.TList(org.apache.thrift.protocol.TType.STRUCT, iprot.readI32()); + struct.new_parts = new ArrayList<PartitionSpec>(_list587.size); + for (int _i588 = 0; _i588 < _list587.size; ++_i588) { - String _elem555; // required - _elem555 = iprot.readString(); - struct.part_vals.add(_elem555); + PartitionSpec _elem589; // required + _elem589 = new PartitionSpec(); + _elem589.read(iprot); + struct.new_parts.add(_elem589); } } - struct.setPart_valsIsSet(true); + struct.setNew_partsIsSet(true); } } } } - public static class append_partition_result implements org.apache.thrift.TBase<append_partition_result, append_partition_result._Fields>, java.io.Serializable, Cloneable { - private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("append_partition_result"); + public static class add_partitions_pspec_result implements org.apache.thrift.TBase<add_partitions_pspec_result, add_partitions_pspec_result._Fields>, java.io.Serializable, Cloneable { + private static final org.apache.thrift.protocol.TStruct STRUCT_DESC = new org.apache.thrift.protocol.TStruct("add_partitions_pspec_result"); - private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.STRUCT, (short)0); + private static final org.apache.thrift.protocol.TField SUCCESS_FIELD_DESC = new org.apache.thrift.protocol.TField("success", org.apache.thrift.protocol.TType.I32, (short)0); private static final org.apache.thrift.protocol.TField O1_FIELD_DESC = new org.apache.thrift.protocol.TField("o1", org.apache.thrift.protocol.TType.STRUCT, (short)1); private static final org.apache.thrift.protocol.TField O2_FIELD_DESC = new org.apache.thrift.protocol.TField("o2", org.apache.thrift.protocol.TType.STRUCT, (short)2); private static final org.apache.thrift.protocol.TField O3_FIELD_DESC = new org.apache.thrift.protocol.TField("o3", org.apache.thrift.protocol.TType.STRUCT, (short)3); private static final Map<Class<? extends IScheme>, SchemeFactory> schemes = new HashMap<Class<? extends IScheme>, SchemeFactory>(); static { - schemes.put(StandardScheme.class, new append_partition_resultStandardSchemeFactory()); - schemes.put(TupleScheme.class, new append_partition_resultTupleSchemeFactory()); + schemes.put(StandardScheme.class, new add_partitions_pspec_resultStandardSchemeFactory()); + schemes.put(TupleScheme.class, new add_partitions_pspec_resultTupleSchemeFactory()); } - private Partition success; // required + private int success; // required private InvalidObjectException o1; // required private AlreadyExistsException o2; // required private MetaException o3; // required @@ -40332,11 +40428,13 @@ public class ThriftHiveMetastore { } // isset id assignments + private static final int __SUCCESS_ISSET_ID = 0; + private byte __isset_bitfield = 0; public static final Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> metaDataMap; static { Map<_Fields, org.apache.thrift.meta_data.FieldMetaData> tmpMap = new EnumMap<_Fields, org.apache.thrift.meta_data.FieldMetaData>(_Fields.class); tmpMap.put(_Fields.SUCCESS, new org.apache.thrift.meta_data.FieldMetaData("success", org.apache.thrift.TFieldRequirementType.DEFAULT, - new org.apache.thrift.meta_data.StructMetaData(org.apache.thrift.protocol.TType.STRUCT, Partition.class))); + new org.apache.thrift.meta_data.FieldValueMetaData(org.apache.thrift.protocol.TType.I32))); tmpMap.put(_Fields.O1, new org.apache.thrift.meta_data.FieldMetaData("o1", org.apache.thrift.TFieldRequirementType.DEFAULT,
[... 23007 lines stripped ...]
