http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java deleted file mode 100644 index b8b7030..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiOrBuilder.java +++ /dev/null @@ -1,258 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/protobuf/api.proto - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -public interface ApiOrBuilder extends - // @@protoc_insertion_point(interface_extends:google.protobuf.Api) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - - /** - * <pre> - * The fully qualified name of this api, including package name - * followed by the api's simple name. - * </pre> - * - * <code>string name = 1;</code> - */ - java.lang.String getName(); - /** - * <pre> - * The fully qualified name of this api, including package name - * followed by the api's simple name. - * </pre> - * - * <code>string name = 1;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString - getNameBytes(); - - /** - * <pre> - * The methods of this api, in unspecified order. - * </pre> - * - * <code>repeated .google.protobuf.Method methods = 2;</code> - */ - java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Method> - getMethodsList(); - /** - * <pre> - * The methods of this api, in unspecified order. - * </pre> - * - * <code>repeated .google.protobuf.Method methods = 2;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.Method getMethods(int index); - /** - * <pre> - * The methods of this api, in unspecified order. - * </pre> - * - * <code>repeated .google.protobuf.Method methods = 2;</code> - */ - int getMethodsCount(); - /** - * <pre> - * The methods of this api, in unspecified order. - * </pre> - * - * <code>repeated .google.protobuf.Method methods = 2;</code> - */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder> - getMethodsOrBuilderList(); - /** - * <pre> - * The methods of this api, in unspecified order. - * </pre> - * - * <code>repeated .google.protobuf.Method methods = 2;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.MethodOrBuilder getMethodsOrBuilder( - int index); - - /** - * <pre> - * Any metadata attached to the API. - * </pre> - * - * <code>repeated .google.protobuf.Option options = 3;</code> - */ - java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Option> - getOptionsList(); - /** - * <pre> - * Any metadata attached to the API. - * </pre> - * - * <code>repeated .google.protobuf.Option options = 3;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.Option getOptions(int index); - /** - * <pre> - * Any metadata attached to the API. - * </pre> - * - * <code>repeated .google.protobuf.Option options = 3;</code> - */ - int getOptionsCount(); - /** - * <pre> - * Any metadata attached to the API. - * </pre> - * - * <code>repeated .google.protobuf.Option options = 3;</code> - */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder> - getOptionsOrBuilderList(); - /** - * <pre> - * Any metadata attached to the API. - * </pre> - * - * <code>repeated .google.protobuf.Option options = 3;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.OptionOrBuilder getOptionsOrBuilder( - int index); - - /** - * <pre> - * A version string for this api. If specified, must have the form - * `major-version.minor-version`, as in `1.10`. If the minor version - * is omitted, it defaults to zero. If the entire version field is - * empty, the major version is derived from the package name, as - * outlined below. If the field is not empty, the version in the - * package name will be verified to be consistent with what is - * provided here. - * The versioning schema uses [semantic - * versioning](http://semver.org) where the major version number - * indicates a breaking change and the minor version an additive, - * non-breaking change. Both version numbers are signals to users - * what to expect from different versions, and should be carefully - * chosen based on the product plan. - * The major version is also reflected in the package name of the - * API, which must end in `v<major-version>`, as in - * `google.feature.v1`. For major versions 0 and 1, the suffix can - * be omitted. Zero major versions must only be used for - * experimental, none-GA apis. - * </pre> - * - * <code>string version = 4;</code> - */ - java.lang.String getVersion(); - /** - * <pre> - * A version string for this api. If specified, must have the form - * `major-version.minor-version`, as in `1.10`. If the minor version - * is omitted, it defaults to zero. If the entire version field is - * empty, the major version is derived from the package name, as - * outlined below. If the field is not empty, the version in the - * package name will be verified to be consistent with what is - * provided here. - * The versioning schema uses [semantic - * versioning](http://semver.org) where the major version number - * indicates a breaking change and the minor version an additive, - * non-breaking change. Both version numbers are signals to users - * what to expect from different versions, and should be carefully - * chosen based on the product plan. - * The major version is also reflected in the package name of the - * API, which must end in `v<major-version>`, as in - * `google.feature.v1`. For major versions 0 and 1, the suffix can - * be omitted. Zero major versions must only be used for - * experimental, none-GA apis. - * </pre> - * - * <code>string version = 4;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString - getVersionBytes(); - - /** - * <pre> - * Source context for the protocol buffer service represented by this - * message. - * </pre> - * - * <code>.google.protobuf.SourceContext source_context = 5;</code> - */ - boolean hasSourceContext(); - /** - * <pre> - * Source context for the protocol buffer service represented by this - * message. - * </pre> - * - * <code>.google.protobuf.SourceContext source_context = 5;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContext getSourceContext(); - /** - * <pre> - * Source context for the protocol buffer service represented by this - * message. - * </pre> - * - * <code>.google.protobuf.SourceContext source_context = 5;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextOrBuilder getSourceContextOrBuilder(); - - /** - * <pre> - * Included APIs. See [Mixin][]. - * </pre> - * - * <code>repeated .google.protobuf.Mixin mixins = 6;</code> - */ - java.util.List<org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin> - getMixinsList(); - /** - * <pre> - * Included APIs. See [Mixin][]. - * </pre> - * - * <code>repeated .google.protobuf.Mixin mixins = 6;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.Mixin getMixins(int index); - /** - * <pre> - * Included APIs. See [Mixin][]. - * </pre> - * - * <code>repeated .google.protobuf.Mixin mixins = 6;</code> - */ - int getMixinsCount(); - /** - * <pre> - * Included APIs. See [Mixin][]. - * </pre> - * - * <code>repeated .google.protobuf.Mixin mixins = 6;</code> - */ - java.util.List<? extends org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder> - getMixinsOrBuilderList(); - /** - * <pre> - * Included APIs. See [Mixin][]. - * </pre> - * - * <code>repeated .google.protobuf.Mixin mixins = 6;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.MixinOrBuilder getMixinsOrBuilder( - int index); - - /** - * <pre> - * The source syntax of the service. - * </pre> - * - * <code>.google.protobuf.Syntax syntax = 7;</code> - */ - int getSyntaxValue(); - /** - * <pre> - * The source syntax of the service. - * </pre> - * - * <code>.google.protobuf.Syntax syntax = 7;</code> - */ - org.apache.hadoop.hbase.shaded.com.google.protobuf.Syntax getSyntax(); -}
http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java deleted file mode 100644 index 817e07d..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ApiProto.java +++ /dev/null @@ -1,98 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/protobuf/api.proto - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -public final class ApiProto { - private ApiProto() {} - public static void registerAllExtensions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite registry) { - } - - public static void registerAllExtensions( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry registry) { - registerAllExtensions( - (org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite) registry); - } - static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_google_protobuf_Api_descriptor; - static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_google_protobuf_Api_fieldAccessorTable; - static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_google_protobuf_Method_descriptor; - static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_google_protobuf_Method_fieldAccessorTable; - static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - internal_static_google_protobuf_Mixin_descriptor; - static final - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internal_static_google_protobuf_Mixin_fieldAccessorTable; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor - getDescriptor() { - return descriptor; - } - private static org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor - descriptor; - static { - java.lang.String[] descriptorData = { - "\n\031google/protobuf/api.proto\022\017google.prot" + - "obuf\032$google/protobuf/source_context.pro" + - "to\032\032google/protobuf/type.proto\"\201\002\n\003Api\022\014" + - "\n\004name\030\001 \001(\t\022(\n\007methods\030\002 \003(\0132\027.google.p" + - "rotobuf.Method\022(\n\007options\030\003 \003(\0132\027.google" + - ".protobuf.Option\022\017\n\007version\030\004 \001(\t\0226\n\016sou" + - "rce_context\030\005 \001(\0132\036.google.protobuf.Sour" + - "ceContext\022&\n\006mixins\030\006 \003(\0132\026.google.proto" + - "buf.Mixin\022\'\n\006syntax\030\007 \001(\0162\027.google.proto" + - "buf.Syntax\"\325\001\n\006Method\022\014\n\004name\030\001 \001(\t\022\030\n\020r", - "equest_type_url\030\002 \001(\t\022\031\n\021request_streami" + - "ng\030\003 \001(\010\022\031\n\021response_type_url\030\004 \001(\t\022\032\n\022r" + - "esponse_streaming\030\005 \001(\010\022(\n\007options\030\006 \003(\013" + - "2\027.google.protobuf.Option\022\'\n\006syntax\030\007 \001(" + - "\0162\027.google.protobuf.Syntax\"#\n\005Mixin\022\014\n\004n" + - "ame\030\001 \001(\t\022\014\n\004root\030\002 \001(\tBu\n\023com.google.pr" + - "otobufB\010ApiProtoP\001Z+google.golang.org/ge" + - "nproto/protobuf/api;api\242\002\003GPB\252\002\036Google.P" + - "rotobuf.WellKnownTypesb\006proto3" - }; - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner = - new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor. InternalDescriptorAssigner() { - public org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistry assignDescriptors( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor root) { - descriptor = root; - return null; - } - }; - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor - .internalBuildGeneratedFileFrom(descriptorData, - new org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FileDescriptor[] { - org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor(), - org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor(), - }, assigner); - internal_static_google_protobuf_Api_descriptor = - getDescriptor().getMessageTypes().get(0); - internal_static_google_protobuf_Api_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_google_protobuf_Api_descriptor, - new java.lang.String[] { "Name", "Methods", "Options", "Version", "SourceContext", "Mixins", "Syntax", }); - internal_static_google_protobuf_Method_descriptor = - getDescriptor().getMessageTypes().get(1); - internal_static_google_protobuf_Method_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_google_protobuf_Method_descriptor, - new java.lang.String[] { "Name", "RequestTypeUrl", "RequestStreaming", "ResponseTypeUrl", "ResponseStreaming", "Options", "Syntax", }); - internal_static_google_protobuf_Mixin_descriptor = - getDescriptor().getMessageTypes().get(2); - internal_static_google_protobuf_Mixin_fieldAccessorTable = new - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable( - internal_static_google_protobuf_Mixin_descriptor, - new java.lang.String[] { "Name", "Root", }); - org.apache.hadoop.hbase.shaded.com.google.protobuf.SourceContextProto.getDescriptor(); - org.apache.hadoop.hbase.shaded.com.google.protobuf.TypeProto.getDescriptor(); - } - - // @@protoc_insertion_point(outer_class_scope) -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java deleted file mode 100644 index bd2f2cd..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingRpcChannel.java +++ /dev/null @@ -1,51 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -/** - * <p>Abstract interface for a blocking RPC channel. {@code BlockingRpcChannel} - * is the blocking equivalent to {@link RpcChannel}. - * - * @author ken...@google.com Kenton Varda - * @author cpov...@google.com Chris Povirk - */ -public interface BlockingRpcChannel { - /** - * Call the given method of the remote service and blocks until it returns. - * {@code callBlockingMethod()} is the blocking equivalent to - * {@link RpcChannel#callMethod}. - */ - Message callBlockingMethod( - Descriptors.MethodDescriptor method, - RpcController controller, - Message request, - Message responsePrototype) throws ServiceException; -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java deleted file mode 100644 index f496f2d..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BlockingService.java +++ /dev/null @@ -1,64 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -/** - * Blocking equivalent to {@link Service}. - * - * @author ken...@google.com Kenton Varda - * @author cpov...@google.com Chris Povirk - */ -public interface BlockingService { - /** - * Equivalent to {@link Service#getDescriptorForType}. - */ - Descriptors.ServiceDescriptor getDescriptorForType(); - - /** - * Equivalent to {@link Service#callMethod}, except that - * {@code callBlockingMethod()} returns the result of the RPC or throws a - * {@link ServiceException} if there is a failure, rather than passing the - * information to a callback. - */ - Message callBlockingMethod(Descriptors.MethodDescriptor method, - RpcController controller, - Message request) throws ServiceException; - - /** - * Equivalent to {@link Service#getRequestPrototype}. - */ - Message getRequestPrototype(Descriptors.MethodDescriptor method); - - /** - * Equivalent to {@link Service#getResponsePrototype}. - */ - Message getResponsePrototype(Descriptors.MethodDescriptor method); -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java deleted file mode 100644 index a2dbe4a..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValue.java +++ /dev/null @@ -1,452 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/protobuf/wrappers.proto - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -/** - * <pre> - * Wrapper message for `bool`. - * The JSON representation for `BoolValue` is JSON `true` and `false`. - * </pre> - * - * Protobuf type {@code google.protobuf.BoolValue} - */ -public final class BoolValue extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 implements - // @@protoc_insertion_point(message_implements:google.protobuf.BoolValue) - BoolValueOrBuilder { - // Use BoolValue.newBuilder() to construct. - private BoolValue(org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<?> builder) { - super(builder); - } - private BoolValue() { - value_ = false; - } - - @java.lang.Override - public final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet - getUnknownFields() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet.getDefaultInstance(); - } - private BoolValue( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - this(); - int mutable_bitField0_ = 0; - try { - boolean done = false; - while (!done) { - int tag = input.readTag(); - switch (tag) { - case 0: - done = true; - break; - default: { - if (!input.skipField(tag)) { - done = true; - } - break; - } - case 8: { - - value_ = input.readBool(); - break; - } - } - } - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - throw e.setUnfinishedMessage(this); - } catch (java.io.IOException e) { - throw new org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException( - e).setUnfinishedMessage(this); - } finally { - makeExtensionsImmutable(); - } - } - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class); - } - - public static final int VALUE_FIELD_NUMBER = 1; - private boolean value_; - /** - * <pre> - * The bool value. - * </pre> - * - * <code>bool value = 1;</code> - */ - public boolean getValue() { - return value_; - } - - private byte memoizedIsInitialized = -1; - public final boolean isInitialized() { - byte isInitialized = memoizedIsInitialized; - if (isInitialized == 1) return true; - if (isInitialized == 0) return false; - - memoizedIsInitialized = 1; - return true; - } - - public void writeTo(org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream output) - throws java.io.IOException { - if (value_ != false) { - output.writeBool(1, value_); - } - } - - public int getSerializedSize() { - int size = memoizedSize; - if (size != -1) return size; - - size = 0; - if (value_ != false) { - size += org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedOutputStream - .computeBoolSize(1, value_); - } - memoizedSize = size; - return size; - } - - private static final long serialVersionUID = 0L; - @java.lang.Override - public boolean equals(final java.lang.Object obj) { - if (obj == this) { - return true; - } - if (!(obj instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)) { - return super.equals(obj); - } - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) obj; - - boolean result = true; - result = result && (getValue() - == other.getValue()); - return result; - } - - @java.lang.Override - public int hashCode() { - if (memoizedHashCode != 0) { - return memoizedHashCode; - } - int hash = 41; - hash = (19 * hash) + getDescriptor().hashCode(); - hash = (37 * hash) + VALUE_FIELD_NUMBER; - hash = (53 * hash) + org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.hashBoolean( - getValue()); - hash = (29 * hash) + unknownFields.hashCode(); - memoizedHashCode = hash; - return hash; - } - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.ByteString data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(byte[] data) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - byte[] data, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return PARSER.parseFrom(data, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom(java.io.InputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseDelimitedFrom( - java.io.InputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseDelimitedWithIOException(PARSER, input, extensionRegistry); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input); - } - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parseFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .parseWithIOException(PARSER, input, extensionRegistry); - } - - public Builder newBuilderForType() { return newBuilder(); } - public static Builder newBuilder() { - return DEFAULT_INSTANCE.toBuilder(); - } - public static Builder newBuilder(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue prototype) { - return DEFAULT_INSTANCE.toBuilder().mergeFrom(prototype); - } - public Builder toBuilder() { - return this == DEFAULT_INSTANCE - ? new Builder() : new Builder().mergeFrom(this); - } - - @java.lang.Override - protected Builder newBuilderForType( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - Builder builder = new Builder(parent); - return builder; - } - /** - * <pre> - * Wrapper message for `bool`. - * The JSON representation for `BoolValue` is JSON `true` and `false`. - * </pre> - * - * Protobuf type {@code google.protobuf.BoolValue} - */ - public static final class Builder extends - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.Builder<Builder> implements - // @@protoc_insertion_point(builder_implements:google.protobuf.BoolValue) - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValueOrBuilder { - public static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptor() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor; - } - - protected org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.FieldAccessorTable - internalGetFieldAccessorTable() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_fieldAccessorTable - .ensureFieldAccessorsInitialized( - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.class, org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.Builder.class); - } - - // Construct using org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.newBuilder() - private Builder() { - maybeForceBuilderInitialization(); - } - - private Builder( - org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3.BuilderParent parent) { - super(parent); - maybeForceBuilderInitialization(); - } - private void maybeForceBuilderInitialization() { - if (org.apache.hadoop.hbase.shaded.com.google.protobuf.GeneratedMessageV3 - .alwaysUseFieldBuilders) { - } - } - public Builder clear() { - super.clear(); - value_ = false; - - return this; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.Descriptor - getDescriptorForType() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.WrappersProto.internal_static_google_protobuf_BoolValue_descriptor; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() { - return org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance(); - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue build() { - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = buildPartial(); - if (!result.isInitialized()) { - throw newUninitializedMessageException(result); - } - return result; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue buildPartial() { - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue result = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue(this); - result.value_ = value_; - onBuilt(); - return result; - } - - public Builder clone() { - return (Builder) super.clone(); - } - public Builder setField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.setField(field, value); - } - public Builder clearField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field) { - return (Builder) super.clearField(field); - } - public Builder clearOneof( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.OneofDescriptor oneof) { - return (Builder) super.clearOneof(oneof); - } - public Builder setRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - int index, Object value) { - return (Builder) super.setRepeatedField(field, index, value); - } - public Builder addRepeatedField( - org.apache.hadoop.hbase.shaded.com.google.protobuf.Descriptors.FieldDescriptor field, - Object value) { - return (Builder) super.addRepeatedField(field, value); - } - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.Message other) { - if (other instanceof org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) { - return mergeFrom((org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue)other); - } else { - super.mergeFrom(other); - return this; - } - } - - public Builder mergeFrom(org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue other) { - if (other == org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue.getDefaultInstance()) return this; - if (other.getValue() != false) { - setValue(other.getValue()); - } - onChanged(); - return this; - } - - public final boolean isInitialized() { - return true; - } - - public Builder mergeFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws java.io.IOException { - org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue parsedMessage = null; - try { - parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry); - } catch (org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException e) { - parsedMessage = (org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue) e.getUnfinishedMessage(); - throw e.unwrapIOException(); - } finally { - if (parsedMessage != null) { - mergeFrom(parsedMessage); - } - } - return this; - } - - private boolean value_ ; - /** - * <pre> - * The bool value. - * </pre> - * - * <code>bool value = 1;</code> - */ - public boolean getValue() { - return value_; - } - /** - * <pre> - * The bool value. - * </pre> - * - * <code>bool value = 1;</code> - */ - public Builder setValue(boolean value) { - - value_ = value; - onChanged(); - return this; - } - /** - * <pre> - * The bool value. - * </pre> - * - * <code>bool value = 1;</code> - */ - public Builder clearValue() { - - value_ = false; - onChanged(); - return this; - } - public final Builder setUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return this; - } - - public final Builder mergeUnknownFields( - final org.apache.hadoop.hbase.shaded.com.google.protobuf.UnknownFieldSet unknownFields) { - return this; - } - - - // @@protoc_insertion_point(builder_scope:google.protobuf.BoolValue) - } - - // @@protoc_insertion_point(class_scope:google.protobuf.BoolValue) - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue DEFAULT_INSTANCE; - static { - DEFAULT_INSTANCE = new org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue(); - } - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstance() { - return DEFAULT_INSTANCE; - } - - private static final org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> - PARSER = new org.apache.hadoop.hbase.shaded.com.google.protobuf.AbstractParser<BoolValue>() { - public BoolValue parsePartialFrom( - org.apache.hadoop.hbase.shaded.com.google.protobuf.CodedInputStream input, - org.apache.hadoop.hbase.shaded.com.google.protobuf.ExtensionRegistryLite extensionRegistry) - throws org.apache.hadoop.hbase.shaded.com.google.protobuf.InvalidProtocolBufferException { - return new BoolValue(input, extensionRegistry); - } - }; - - public static org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> parser() { - return PARSER; - } - - @java.lang.Override - public org.apache.hadoop.hbase.shaded.com.google.protobuf.Parser<BoolValue> getParserForType() { - return PARSER; - } - - public org.apache.hadoop.hbase.shaded.com.google.protobuf.BoolValue getDefaultInstanceForType() { - return DEFAULT_INSTANCE; - } - -} - http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java deleted file mode 100644 index c1d3b5f..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BoolValueOrBuilder.java +++ /dev/null @@ -1,18 +0,0 @@ -// Generated by the protocol buffer compiler. DO NOT EDIT! -// source: google/protobuf/wrappers.proto - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -public interface BoolValueOrBuilder extends - // @@protoc_insertion_point(interface_extends:google.protobuf.BoolValue) - org.apache.hadoop.hbase.shaded.com.google.protobuf.MessageOrBuilder { - - /** - * <pre> - * The bool value. - * </pre> - * - * <code>bool value = 1;</code> - */ - boolean getValue(); -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java deleted file mode 100644 index 6bc9641..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/BooleanArrayList.java +++ /dev/null @@ -1,272 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -import org.apache.hadoop.hbase.shaded.com.google.protobuf.Internal.BooleanList; - -import java.util.Arrays; -import java.util.Collection; -import java.util.RandomAccess; - -/** - * An implementation of {@link BooleanList} on top of a primitive array. - * - * @author dw...@google.com (Daniel Weis) - */ -final class BooleanArrayList - extends AbstractProtobufList<Boolean> - implements BooleanList, RandomAccess { - - private static final BooleanArrayList EMPTY_LIST = new BooleanArrayList(); - static { - EMPTY_LIST.makeImmutable(); - } - - public static BooleanArrayList emptyList() { - return EMPTY_LIST; - } - - /** - * The backing store for the list. - */ - private boolean[] array; - - /** - * The size of the list distinct from the length of the array. That is, it is the number of - * elements set in the list. - */ - private int size; - - /** - * Constructs a new mutable {@code BooleanArrayList} with default capacity. - */ - BooleanArrayList() { - this(new boolean[DEFAULT_CAPACITY], 0); - } - - /** - * Constructs a new mutable {@code BooleanArrayList} - * containing the same elements as {@code other}. - */ - private BooleanArrayList(boolean[] other, int size) { - array = other; - this.size = size; - } - - @Override - public boolean equals(Object o) { - if (this == o) { - return true; - } - if (!(o instanceof BooleanArrayList)) { - return super.equals(o); - } - BooleanArrayList other = (BooleanArrayList) o; - if (size != other.size) { - return false; - } - - final boolean[] arr = other.array; - for (int i = 0; i < size; i++) { - if (array[i] != arr[i]) { - return false; - } - } - - return true; - } - - @Override - public int hashCode() { - int result = 1; - for (int i = 0; i < size; i++) { - result = (31 * result) + Internal.hashBoolean(array[i]); - } - return result; - } - - @Override - public BooleanList mutableCopyWithCapacity(int capacity) { - if (capacity < size) { - throw new IllegalArgumentException(); - } - return new BooleanArrayList(Arrays.copyOf(array, capacity), size); - } - - @Override - public Boolean get(int index) { - return getBoolean(index); - } - - @Override - public boolean getBoolean(int index) { - ensureIndexInRange(index); - return array[index]; - } - - @Override - public int size() { - return size; - } - - @Override - public Boolean set(int index, Boolean element) { - return setBoolean(index, element); - } - - @Override - public boolean setBoolean(int index, boolean element) { - ensureIsMutable(); - ensureIndexInRange(index); - boolean previousValue = array[index]; - array[index] = element; - return previousValue; - } - - @Override - public void add(int index, Boolean element) { - addBoolean(index, element); - } - - /** - * Like {@link #add(Boolean)} but more efficient in that it doesn't box the element. - */ - @Override - public void addBoolean(boolean element) { - addBoolean(size, element); - } - - /** - * Like {@link #add(int, Boolean)} but more efficient in that it doesn't box the element. - */ - private void addBoolean(int index, boolean element) { - ensureIsMutable(); - if (index < 0 || index > size) { - throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index)); - } - - if (size < array.length) { - // Shift everything over to make room - System.arraycopy(array, index, array, index + 1, size - index); - } else { - // Resize to 1.5x the size - int length = ((size * 3) / 2) + 1; - boolean[] newArray = new boolean[length]; - - // Copy the first part directly - System.arraycopy(array, 0, newArray, 0, index); - - // Copy the rest shifted over by one to make room - System.arraycopy(array, index, newArray, index + 1, size - index); - array = newArray; - } - - array[index] = element; - size++; - modCount++; - } - - @Override - public boolean addAll(Collection<? extends Boolean> collection) { - ensureIsMutable(); - - if (collection == null) { - throw new NullPointerException(); - } - - // We specialize when adding another BooleanArrayList to avoid boxing elements. - if (!(collection instanceof BooleanArrayList)) { - return super.addAll(collection); - } - - BooleanArrayList list = (BooleanArrayList) collection; - if (list.size == 0) { - return false; - } - - int overflow = Integer.MAX_VALUE - size; - if (overflow < list.size) { - // We can't actually represent a list this large. - throw new OutOfMemoryError(); - } - - int newSize = size + list.size; - if (newSize > array.length) { - array = Arrays.copyOf(array, newSize); - } - - System.arraycopy(list.array, 0, array, size, list.size); - size = newSize; - modCount++; - return true; - } - - @Override - public boolean remove(Object o) { - ensureIsMutable(); - for (int i = 0; i < size; i++) { - if (o.equals(array[i])) { - System.arraycopy(array, i + 1, array, i, size - i); - size--; - modCount++; - return true; - } - } - return false; - } - - @Override - public Boolean remove(int index) { - ensureIsMutable(); - ensureIndexInRange(index); - boolean value = array[index]; - System.arraycopy(array, index + 1, array, index, size - index); - size--; - modCount++; - return value; - } - - /** - * Ensures that the provided {@code index} is within the range of {@code [0, size]}. Throws an - * {@link IndexOutOfBoundsException} if it is not. - * - * @param index the index to verify is in range - */ - private void ensureIndexInRange(int index) { - if (index < 0 || index >= size) { - throw new IndexOutOfBoundsException(makeOutOfBoundsExceptionMessage(index)); - } - } - - private String makeOutOfBoundsExceptionMessage(int index) { - return "Index:" + index + ", Size:" + size; - } -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java deleted file mode 100644 index 553243a..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteBufferWriter.java +++ /dev/null @@ -1,185 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -import static java.lang.Math.max; -import static java.lang.Math.min; - -import java.io.IOException; -import java.io.OutputStream; -import java.lang.ref.SoftReference; -import java.lang.reflect.Field; -import java.nio.ByteBuffer; -import java.nio.channels.WritableByteChannel; - -/** - * Utility class to provide efficient writing of {@link ByteBuffer}s to {@link OutputStream}s. - */ -final class ByteBufferWriter { - private ByteBufferWriter() {} - - /** - * Minimum size for a cached buffer. This prevents us from allocating buffers that are too - * small to be easily reused. - */ - // TODO(nathanmittler): tune this property or allow configuration? - private static final int MIN_CACHED_BUFFER_SIZE = 1024; - - /** - * Maximum size for a cached buffer. If a larger buffer is required, it will be allocated - * but not cached. - */ - // TODO(nathanmittler): tune this property or allow configuration? - private static final int MAX_CACHED_BUFFER_SIZE = 16 * 1024; - - /** - * The fraction of the requested buffer size under which the buffer will be reallocated. - */ - // TODO(nathanmittler): tune this property or allow configuration? - private static final float BUFFER_REALLOCATION_THRESHOLD = 0.5f; - - /** - * Keeping a soft reference to a thread-local buffer. This buffer is used for writing a - * {@link ByteBuffer} to an {@link OutputStream} when no zero-copy alternative was available. - * Using a "soft" reference since VMs may keep this reference around longer than "weak" - * (e.g. HotSpot will maintain soft references until memory pressure warrants collection). - */ - private static final ThreadLocal<SoftReference<byte[]>> BUFFER = - new ThreadLocal<SoftReference<byte[]>>(); - - /** - * This is a hack for GAE, where {@code FileOutputStream} is unavailable. - */ - private static final Class<?> FILE_OUTPUT_STREAM_CLASS = safeGetClass("java.io.FileOutputStream"); - private static final long CHANNEL_FIELD_OFFSET = getChannelFieldOffset(FILE_OUTPUT_STREAM_CLASS); - - /** - * For testing purposes only. Clears the cached buffer to force a new allocation on the next - * invocation. - */ - static void clearCachedBuffer() { - BUFFER.set(null); - } - - /** - * Writes the remaining content of the buffer to the given stream. The buffer {@code position} - * will remain unchanged by this method. - */ - static void write(ByteBuffer buffer, OutputStream output) throws IOException { - final int initialPos = buffer.position(); - try { - if (buffer.hasArray()) { - // Optimized write for array-backed buffers. - // Note that we're taking the risk that a malicious OutputStream could modify the array. - output.write(buffer.array(), buffer.arrayOffset() + buffer.position(), buffer.remaining()); - } else if (!writeToChannel(buffer, output)){ - // Read all of the data from the buffer to an array. - // TODO(nathanmittler): Consider performance improvements for other "known" stream types. - final byte[] array = getOrCreateBuffer(buffer.remaining()); - while (buffer.hasRemaining()) { - int length = min(buffer.remaining(), array.length); - buffer.get(array, 0, length); - output.write(array, 0, length); - } - } - } finally { - // Restore the initial position. - buffer.position(initialPos); - } - } - - static byte[] getOrCreateBuffer(int requestedSize) { - requestedSize = max(requestedSize, MIN_CACHED_BUFFER_SIZE); - - byte[] buffer = getBuffer(); - // Only allocate if we need to. - if (buffer == null || needToReallocate(requestedSize, buffer.length)) { - buffer = new byte[requestedSize]; - - // Only cache the buffer if it's not too big. - if (requestedSize <= MAX_CACHED_BUFFER_SIZE) { - setBuffer(buffer); - } - } - return buffer; - } - - private static boolean needToReallocate(int requestedSize, int bufferLength) { - // First check against just the requested length to avoid the multiply. - return bufferLength < requestedSize - && bufferLength < requestedSize * BUFFER_REALLOCATION_THRESHOLD; - } - - private static byte[] getBuffer() { - SoftReference<byte[]> sr = BUFFER.get(); - return sr == null ? null : sr.get(); - } - - private static void setBuffer(byte[] value) { - BUFFER.set(new SoftReference<byte[]>(value)); - } - - private static boolean writeToChannel(ByteBuffer buffer, OutputStream output) throws IOException { - if (CHANNEL_FIELD_OFFSET >= 0 && FILE_OUTPUT_STREAM_CLASS.isInstance(output)) { - // Use a channel to write out the ByteBuffer. This will automatically empty the buffer. - WritableByteChannel channel = null; - try { - channel = (WritableByteChannel) UnsafeUtil.getObject(output, CHANNEL_FIELD_OFFSET); - } catch (ClassCastException e) { - // Absorb. - } - if (channel != null) { - channel.write(buffer); - return true; - } - } - return false; - } - - private static Class<?> safeGetClass(String className) { - try { - return Class.forName(className); - } catch (ClassNotFoundException e) { - return null; - } - } - private static long getChannelFieldOffset(Class<?> clazz) { - try { - if (clazz != null && UnsafeUtil.hasUnsafeArrayOperations()) { - Field field = clazz.getDeclaredField("channel"); - return UnsafeUtil.objectFieldOffset(field); - } - } catch (Throwable e) { - // Absorb - } - return -1; - } -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java deleted file mode 100644 index a745d37..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInput.java +++ /dev/null @@ -1,81 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -import java.io.IOException; -import java.nio.ByteBuffer; - -/** - * An input for raw bytes. This is similar to an InputStream but it is offset addressable. All the - * read APIs are relative. - */ -@ExperimentalApi -public abstract class ByteInput { - - /** - * Reads a single byte from the given offset. - * @param offset The offset from where byte to be read - * @return The byte of data at given offset - */ - public abstract byte read(int offset); - - /** - * Reads bytes of data from the given offset into an array of bytes. - * @param offset The src offset within this ByteInput from where data to be read. - * @param out Destination byte array to read data into. - * @return The number of bytes read from ByteInput - */ - public int read(int offset, byte b[]) throws IOException { - return read(offset, b, 0, b.length); - } - - /** - * Reads up to <code>len</code> bytes of data from the given offset into an array of bytes. - * @param offset The src offset within this ByteInput from where data to be read. - * @param out Destination byte array to read data into. - * @param outOffset Offset within the the out byte[] where data to be read into. - * @param len The number of bytes to read. - * @return The number of bytes read from ByteInput - */ - public abstract int read(int offset, byte[] out, int outOffset, int len); - - /** - * Reads bytes of data from the given offset into given {@link ByteBuffer}. - * @param offset he src offset within this ByteInput from where data to be read. - * @param out Destination {@link ByteBuffer} to read data into. - * @return The number of bytes read from ByteInput - */ - public abstract int read(int offset, ByteBuffer out); - - /** - * @return Total number of bytes in this ByteInput. - */ - public abstract int size(); -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java deleted file mode 100644 index 30de4ec..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteInputByteString.java +++ /dev/null @@ -1,251 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -import java.io.IOException; -import java.io.InputStream; -import java.io.InvalidObjectException; -import java.io.ObjectInputStream; -import java.io.OutputStream; -import java.nio.ByteBuffer; -import java.nio.charset.Charset; -import java.util.Arrays; -import java.util.Collections; -import java.util.List; - -/** - * A {@link ByteString} that wraps around a {@link ByteInput}. - */ -final class ByteInputByteString extends ByteString.LeafByteString { - private final ByteInput buffer; - private final int offset, length; - - ByteInputByteString(ByteInput buffer, int offset, int length) { - if (buffer == null) { - throw new NullPointerException("buffer"); - } - this.buffer = buffer; - this.offset = offset; - this.length = length; - } - - // ================================================================= - // Serializable - - /** - * Magic method that lets us override serialization behavior. - */ - private Object writeReplace() { - return ByteString.wrap(toByteArray()); - } - - /** - * Magic method that lets us override deserialization behavior. - */ - private void readObject(@SuppressWarnings("unused") ObjectInputStream in) throws IOException { - throw new InvalidObjectException("ByteInputByteString instances are not to be serialized directly");// TODO check here - } - - // ================================================================= - - @Override - public byte byteAt(int index) { - return buffer.read(getAbsoluteOffset(index)); - } - - private int getAbsoluteOffset(int relativeOffset) { - return this.offset + relativeOffset; - } - - @Override - public int size() { - return length; - } - - @Override - public ByteString substring(int beginIndex, int endIndex) { - if (beginIndex < 0 || beginIndex >= size() || endIndex < beginIndex || endIndex >= size()) { - throw new IllegalArgumentException( - String.format("Invalid indices [%d, %d]", beginIndex, endIndex)); - } - return new ByteInputByteString(this.buffer, getAbsoluteOffset(beginIndex), endIndex - beginIndex); - } - - @Override - protected void copyToInternal( - byte[] target, int sourceOffset, int targetOffset, int numberToCopy) { - this.buffer.read(getAbsoluteOffset(sourceOffset), target, targetOffset, numberToCopy); - } - - @Override - public void copyTo(ByteBuffer target) { - this.buffer.read(this.offset, target); - } - - @Override - public void writeTo(OutputStream out) throws IOException { - out.write(toByteArray());// TODO - } - - @Override - boolean equalsRange(ByteString other, int offset, int length) { - return substring(0, length).equals(other.substring(offset, offset + length)); - } - - @Override - void writeToInternal(OutputStream out, int sourceOffset, int numberToWrite) throws IOException { - byte[] buf = ByteBufferWriter.getOrCreateBuffer(numberToWrite); - this.buffer.read(getAbsoluteOffset(sourceOffset), buf, 0, numberToWrite); - out.write(buf, 0, numberToWrite); - } - - @Override - void writeTo(ByteOutput output) throws IOException { - output.writeLazy(toByteArray(), 0, length); - } - - @Override - public ByteBuffer asReadOnlyByteBuffer() { - return ByteBuffer.wrap(toByteArray()).asReadOnlyBuffer(); - } - - @Override - public List<ByteBuffer> asReadOnlyByteBufferList() { - return Collections.singletonList(asReadOnlyByteBuffer()); - } - - @Override - protected String toStringInternal(Charset charset) { - byte[] bytes = toByteArray(); - return new String(bytes, 0, bytes.length, charset); - } - - @Override - public boolean isValidUtf8() { - return Utf8.isValidUtf8(buffer, offset, offset + length); - } - - @Override - protected int partialIsValidUtf8(int state, int offset, int length) { - int off = getAbsoluteOffset(offset); - return Utf8.partialIsValidUtf8(state, buffer, off, off + length); - } - - @Override - public boolean equals(Object other) { - if (other == this) { - return true; - } - if (!(other instanceof ByteString)) { - return false; - } - ByteString otherString = ((ByteString) other); - if (size() != otherString.size()) { - return false; - } - if (size() == 0) { - return true; - } - if (other instanceof RopeByteString) { - return other.equals(this); - } - return Arrays.equals(this.toByteArray(), otherString.toByteArray()); - } - - @Override - protected int partialHash(int h, int offset, int length) { - offset = getAbsoluteOffset(offset); - int end = offset + length; - for (int i = offset; i < end; i++) { - h = h * 31 + buffer.read(i); - } - return h; - } - - @Override - public InputStream newInput() { - return new InputStream() { - private final ByteInput buf = buffer; - private int pos = offset; - private int limit = pos + length; - private int mark = pos; - - @Override - public void mark(int readlimit) { - this.mark = readlimit; - } - - @Override - public boolean markSupported() { - return true; - } - - @Override - public void reset() throws IOException { - this.pos = this.mark; - } - - @Override - public int available() throws IOException { - return this.limit - this.pos; - } - - @Override - public int read() throws IOException { - if (available() <= 0) { - return -1; - } - return this.buf.read(pos++) & 0xFF; - } - - @Override - public int read(byte[] bytes, int off, int len) throws IOException { - int remain = available(); - if (remain <= 0) { - return -1; - } - len = Math.min(len, remain); - buf.read(pos, bytes, off, len); - pos += len; - return len; - } - }; - } - - @Override - public CodedInputStream newCodedInput() { - // We trust CodedInputStream not to modify the bytes, or to give anyone - // else access to them. - CodedInputStream cis = CodedInputStream.newInstance(buffer, offset, length, true); - cis.enableAliasing(true); - return cis; - } -} http://git-wip-us.apache.org/repos/asf/hbase/blob/7a6de1bd/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java ---------------------------------------------------------------------- diff --git a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java b/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java deleted file mode 100644 index 8a83fb2..0000000 --- a/hbase-protocol-shaded/src/main/java/org/apache/hadoop/hbase/shaded/com/google/protobuf/ByteOutput.java +++ /dev/null @@ -1,116 +0,0 @@ -// Protocol Buffers - Google's data interchange format -// Copyright 2008 Google Inc. All rights reserved. -// https://developers.google.com/protocol-buffers/ -// -// Redistribution and use in source and binary forms, with or without -// modification, are permitted provided that the following conditions are -// met: -// -// * Redistributions of source code must retain the above copyright -// notice, this list of conditions and the following disclaimer. -// * Redistributions in binary form must reproduce the above -// copyright notice, this list of conditions and the following disclaimer -// in the documentation and/or other materials provided with the -// distribution. -// * Neither the name of Google Inc. nor the names of its -// contributors may be used to endorse or promote products derived from -// this software without specific prior written permission. -// -// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS -// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT -// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR -// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT -// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, -// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT -// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, -// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY -// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT -// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE -// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. - -package org.apache.hadoop.hbase.shaded.com.google.protobuf; - -import java.io.IOException; -import java.nio.ByteBuffer; - -/** - * An output target for raw bytes. This interface provides semantics that support two types of - * writing: - * - * <p><b>Traditional write operations:</b> - * (as defined by {@link java.io.OutputStream}) where the target method is responsible for either - * copying the data or completing the write before returning from the method call. - * - * <p><b>Lazy write operations:</b> where the caller guarantees that it will never modify the - * provided buffer and it can therefore be considered immutable. The target method is free to - * maintain a reference to the buffer beyond the scope of the method call (e.g. until the write - * operation completes). - */ -@ExperimentalApi -public abstract class ByteOutput { - /** - * Writes a single byte. - * - * @param value the byte to be written - * @throws IOException thrown if an error occurred while writing - */ - public abstract void write(byte value) throws IOException; - - /** - * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will - * not be processed prior to the return of this method call, since {@code value} may be - * reused/altered by the caller. - * - * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a - * programming error and will lead to data corruption which will be difficult to debug. - * - * @param value the bytes to be written - * @param offset the offset of the start of the writable range - * @param length the number of bytes to write starting from {@code offset} - * @throws IOException thrown if an error occurred while writing - */ - public abstract void write(byte[] value, int offset, int length) throws IOException; - - /** - * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value - * beyond the scope of this method call (e.g. write later) since it is considered immutable and is - * guaranteed not to change by the caller. - * - * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a - * programming error and will lead to data corruption which will be difficult to debug. - * - * @param value the bytes to be written - * @param offset the offset of the start of the writable range - * @param length the number of bytes to write starting from {@code offset} - * @throws IOException thrown if an error occurred while writing - */ - public abstract void writeLazy(byte[] value, int offset, int length) throws IOException; - - /** - * Writes a sequence of bytes. The {@link ByteOutput} must copy {@code value} if it will - * not be processed prior to the return of this method call, since {@code value} may be - * reused/altered by the caller. - * - * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a - * programming error and will lead to data corruption which will be difficult to debug. - * - * @param value the bytes to be written. Upon returning from this call, the {@code position} of - * this buffer will be set to the {@code limit} - * @throws IOException thrown if an error occurred while writing - */ - public abstract void write(ByteBuffer value) throws IOException; - - /** - * Writes a sequence of bytes. The {@link ByteOutput} is free to retain a reference to the value - * beyond the scope of this method call (e.g. write later) since it is considered immutable and is - * guaranteed not to change by the caller. - * - * <p>NOTE: This method <strong>MUST NOT</strong> modify the {@code value}. Doing so is a - * programming error and will lead to data corruption which will be difficult to debug. - * - * @param value the bytes to be written. Upon returning from this call, the {@code position} of - * this buffer will be set to the {@code limit} - * @throws IOException thrown if an error occurred while writing - */ - public abstract void writeLazy(ByteBuffer value) throws IOException; -}