hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject [02/39] hbase git commit: HBASE-16263 Move all to do w/ protobuf -- *.proto files and generated classes -- under hbase-protocol
Date Thu, 21 Jul 2016 17:20:25 GMT
http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
deleted file mode 100644
index 3fd34e9..0000000
--- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/protobuf/generated/TestRpcServiceProtos.java
+++ /dev/null
@@ -1,424 +0,0 @@
-// Generated by the protocol buffer compiler.  DO NOT EDIT!
-// source: test_rpc_service.proto
-
-package org.apache.hadoop.hbase.ipc.protobuf.generated;
-
-public final class TestRpcServiceProtos {
-  private TestRpcServiceProtos() {}
-  public static void registerAllExtensions(
-      com.google.protobuf.ExtensionRegistry registry) {
-  }
-  /**
-   * Protobuf service {@code TestProtobufRpcProto}
-   *
-   * <pre>
-   **
-   * A protobuf service for use in tests
-   * </pre>
-   */
-  public static abstract class TestProtobufRpcProto
-      implements com.google.protobuf.Service {
-    protected TestProtobufRpcProto() {}
-
-    public interface Interface {
-      /**
-       * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-       */
-      public abstract void ping(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done);
-
-      /**
-       * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
-       */
-      public abstract void echo(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>
done);
-
-      /**
-       * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-       */
-      public abstract void error(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done);
-
-    }
-
-    public static com.google.protobuf.Service newReflectiveService(
-        final Interface impl) {
-      return new TestProtobufRpcProto() {
-        @java.lang.Override
-        public  void ping(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done) {
-          impl.ping(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void echo(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>
done) {
-          impl.echo(controller, request, done);
-        }
-
-        @java.lang.Override
-        public  void error(
-            com.google.protobuf.RpcController controller,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-            com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done) {
-          impl.error(controller, request, done);
-        }
-
-      };
-    }
-
-    public static com.google.protobuf.BlockingService
-        newReflectiveBlockingService(final BlockingInterface impl) {
-      return new com.google.protobuf.BlockingService() {
-        public final com.google.protobuf.Descriptors.ServiceDescriptor
-            getDescriptorForType() {
-          return getDescriptor();
-        }
-
-        public final com.google.protobuf.Message callBlockingMethod(
-            com.google.protobuf.Descriptors.MethodDescriptor method,
-            com.google.protobuf.RpcController controller,
-            com.google.protobuf.Message request)
-            throws com.google.protobuf.ServiceException {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.callBlockingMethod() given method descriptor for " +
-              "wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return impl.ping(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
-            case 1:
-              return impl.echo(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)request);
-            case 2:
-              return impl.error(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request);
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getRequestPrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getRequestPrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-        public final com.google.protobuf.Message
-            getResponsePrototype(
-            com.google.protobuf.Descriptors.MethodDescriptor method) {
-          if (method.getService() != getDescriptor()) {
-            throw new java.lang.IllegalArgumentException(
-              "Service.getResponsePrototype() given method " +
-              "descriptor for wrong service type.");
-          }
-          switch(method.getIndex()) {
-            case 0:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-            case 1:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
-            case 2:
-              return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-            default:
-              throw new java.lang.AssertionError("Can't get here.");
-          }
-        }
-
-      };
-    }
-
-    /**
-     * <code>rpc ping(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-     */
-    public abstract void ping(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done);
-
-    /**
-     * <code>rpc echo(.EchoRequestProto) returns (.EchoResponseProto);</code>
-     */
-    public abstract void echo(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>
done);
-
-    /**
-     * <code>rpc error(.EmptyRequestProto) returns (.EmptyResponseProto);</code>
-     */
-    public abstract void error(
-        com.google.protobuf.RpcController controller,
-        org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-        com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done);
-
-    public static final
-        com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptor() {
-      return org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.getDescriptor().getServices().get(0);
-    }
-    public final com.google.protobuf.Descriptors.ServiceDescriptor
-        getDescriptorForType() {
-      return getDescriptor();
-    }
-
-    public final void callMethod(
-        com.google.protobuf.Descriptors.MethodDescriptor method,
-        com.google.protobuf.RpcController controller,
-        com.google.protobuf.Message request,
-        com.google.protobuf.RpcCallback<
-          com.google.protobuf.Message> done) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.callMethod() given method descriptor for wrong " +
-          "service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          this.ping(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
-              done));
-          return;
-        case 1:
-          this.echo(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>specializeCallback(
-              done));
-          return;
-        case 2:
-          this.error(controller, (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto)request,
-            com.google.protobuf.RpcUtil.<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>specializeCallback(
-              done));
-          return;
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getRequestPrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getRequestPrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public final com.google.protobuf.Message
-        getResponsePrototype(
-        com.google.protobuf.Descriptors.MethodDescriptor method) {
-      if (method.getService() != getDescriptor()) {
-        throw new java.lang.IllegalArgumentException(
-          "Service.getResponsePrototype() given method " +
-          "descriptor for wrong service type.");
-      }
-      switch(method.getIndex()) {
-        case 0:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-        case 1:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance();
-        case 2:
-          return org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance();
-        default:
-          throw new java.lang.AssertionError("Can't get here.");
-      }
-    }
-
-    public static Stub newStub(
-        com.google.protobuf.RpcChannel channel) {
-      return new Stub(channel);
-    }
-
-    public static final class Stub extends org.apache.hadoop.hbase.ipc.protobuf.generated.TestRpcServiceProtos.TestProtobufRpcProto
implements Interface {
-      private Stub(com.google.protobuf.RpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.RpcChannel channel;
-
-      public com.google.protobuf.RpcChannel getChannel() {
-        return channel;
-      }
-
-      public  void ping(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
-      }
-
-      public  void echo(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto>
done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.class,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance()));
-      }
-
-      public  void error(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request,
-          com.google.protobuf.RpcCallback<org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto>
done) {
-        channel.callMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance(),
-          com.google.protobuf.RpcUtil.generalizeCallback(
-            done,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.class,
-            org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance()));
-      }
-    }
-
-    public static BlockingInterface newBlockingStub(
-        com.google.protobuf.BlockingRpcChannel channel) {
-      return new BlockingStub(channel);
-    }
-
-    public interface BlockingInterface {
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto
ping(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto
echo(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
-          throws com.google.protobuf.ServiceException;
-
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto
error(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws com.google.protobuf.ServiceException;
-    }
-
-    private static final class BlockingStub implements BlockingInterface {
-      private BlockingStub(com.google.protobuf.BlockingRpcChannel channel) {
-        this.channel = channel;
-      }
-
-      private final com.google.protobuf.BlockingRpcChannel channel;
-
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto
ping(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto)
channel.callBlockingMethod(
-          getDescriptor().getMethods().get(0),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto
echo(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoRequestProto request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto)
channel.callBlockingMethod(
-          getDescriptor().getMethods().get(1),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EchoResponseProto.getDefaultInstance());
-      }
-
-
-      public org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto
error(
-          com.google.protobuf.RpcController controller,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyRequestProto request)
-          throws com.google.protobuf.ServiceException {
-        return (org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto)
channel.callBlockingMethod(
-          getDescriptor().getMethods().get(2),
-          controller,
-          request,
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.EmptyResponseProto.getDefaultInstance());
-      }
-
-    }
-
-    // @@protoc_insertion_point(class_scope:TestProtobufRpcProto)
-  }
-
-
-  public static com.google.protobuf.Descriptors.FileDescriptor
-      getDescriptor() {
-    return descriptor;
-  }
-  private static com.google.protobuf.Descriptors.FileDescriptor
-      descriptor;
-  static {
-    java.lang.String[] descriptorData = {
-      "\n\026test_rpc_service.proto\032\ntest.proto2\250\001\n" +
-      "\024TestProtobufRpcProto\022/\n\004ping\022\022.EmptyReq" +
-      "uestProto\032\023.EmptyResponseProto\022-\n\004echo\022\021" +
-      ".EchoRequestProto\032\022.EchoResponseProto\0220\n" +
-      "\005error\022\022.EmptyRequestProto\032\023.EmptyRespon" +
-      "seProtoBL\n.org.apache.hadoop.hbase.ipc.p" +
-      "rotobuf.generatedB\024TestRpcServiceProtos\210" +
-      "\001\001\240\001\001"
-    };
-    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
-      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
-        public com.google.protobuf.ExtensionRegistry assignDescriptors(
-            com.google.protobuf.Descriptors.FileDescriptor root) {
-          descriptor = root;
-          return null;
-        }
-      };
-    com.google.protobuf.Descriptors.FileDescriptor
-      .internalBuildGeneratedFileFrom(descriptorData,
-        new com.google.protobuf.Descriptors.FileDescriptor[] {
-          org.apache.hadoop.hbase.ipc.protobuf.generated.TestProtos.getDescriptor(),
-        }, assigner);
-  }
-
-  // @@protoc_insertion_point(outer_class_scope)
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
deleted file mode 100644
index b4dc01e..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationNullResponseProtocol.proto
+++ /dev/null
@@ -1,38 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationWithNullResponseProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-// use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-// protoc or hadoop's protoc compiler.
-message ColumnAggregationNullResponseSumRequest {
-  required bytes family = 1;
-  optional bytes qualifier = 2;
-}
-
-message ColumnAggregationNullResponseSumResponse {
-  optional int64 sum = 1;
-}
-
-service ColumnAggregationServiceNullResponse {
-  rpc sum(ColumnAggregationNullResponseSumRequest)
-    returns(ColumnAggregationNullResponseSumResponse);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
deleted file mode 100644
index ad1acda..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationProtocol.proto
+++ /dev/null
@@ -1,35 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message SumRequest {
-  required bytes family = 1;
-  optional bytes qualifier = 2;
-}
-
-message SumResponse {
-  required int64 sum = 1;
-}
-
-service ColumnAggregationService {
-  rpc sum(SumRequest) returns(SumResponse);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto b/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
deleted file mode 100644
index 7808949..0000000
--- a/hbase-server/src/test/protobuf/ColumnAggregationWithErrorsProtocol.proto
+++ /dev/null
@@ -1,38 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "ColumnAggregationWithErrorsProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-// use unique names for messages in ColumnAggregationXXX.protos due to a bug in
-// protoc or hadoop's protoc compiler.
-message ColumnAggregationWithErrorsSumRequest {
-  required bytes family = 1;
-  optional bytes qualifier = 2;
-}
-
-message ColumnAggregationWithErrorsSumResponse {
-  required int64 sum = 1;
-}
-
-service ColumnAggregationServiceWithErrors {
-  rpc sum(ColumnAggregationWithErrorsSumRequest)
-    returns(ColumnAggregationWithErrorsSumResponse);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto b/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
deleted file mode 100644
index 539f7da..0000000
--- a/hbase-server/src/test/protobuf/DummyRegionServerEndpoint.proto
+++ /dev/null
@@ -1,37 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-package hbase.test.pb;
-
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "DummyRegionServerEndpointProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message DummyRequest {
-}
-
-message DummyResponse {
-  required string value = 1;
-}
-
-service DummyService {
-  rpc dummyCall(DummyRequest) returns(DummyResponse);
-  rpc dummyThrow(DummyRequest) returns(DummyResponse);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto b/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
deleted file mode 100644
index b8c77ca..0000000
--- a/hbase-server/src/test/protobuf/IncrementCounterProcessor.proto
+++ /dev/null
@@ -1,55 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "IncrementCounterProcessorTestProtos";
-option java_generate_equals_and_hash = true;
-
-message IncCounterProcessorRequest {
-  required bytes row = 1;
-  required int32 counter = 2;
-}
-
-message IncCounterProcessorResponse {
-  required int32 response = 1;
-}
-
-message FriendsOfFriendsProcessorRequest {
-  required bytes person = 1;
-  required bytes row = 2;
-  repeated string result = 3;
-}
-
-message FriendsOfFriendsProcessorResponse {
-  repeated string result = 1;
-}
-
-message RowSwapProcessorRequest {
-  required bytes row1 = 1;
-  required bytes row2 = 2;
-}
-
-message RowSwapProcessorResponse {
-}
-
-message TimeoutProcessorRequest {
-  required bytes row = 1;
-}
-
-message TimeoutProcessorResponse {
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/PingProtocol.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/PingProtocol.proto b/hbase-server/src/test/protobuf/PingProtocol.proto
deleted file mode 100644
index ef63ee0..0000000
--- a/hbase-server/src/test/protobuf/PingProtocol.proto
+++ /dev/null
@@ -1,66 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-// Coprocessor test
-option java_package = "org.apache.hadoop.hbase.coprocessor.protobuf.generated";
-option java_outer_classname = "PingProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-message PingRequest {
-}
-
-message PingResponse {
-  required string pong = 1;
-}
-
-message CountRequest {
-}
-
-message CountResponse {
-  required int32 count = 1;
-}
-
-message IncrementCountRequest {
-  required int32 diff = 1;
-}
-
-message IncrementCountResponse {
-  required int32 count = 1;
-}
-
-message HelloRequest {
-  optional string name = 1;
-}
-
-message HelloResponse {
-  optional string response = 1;
-}
-
-message NoopRequest {
-}
-
-message NoopResponse {
-}
-
-service PingService {
-  rpc ping(PingRequest) returns(PingResponse);
-  rpc count(CountRequest) returns(CountResponse);
-  rpc increment(IncrementCountRequest) returns(IncrementCountResponse);
-  rpc hello(HelloRequest) returns(HelloResponse);
-  rpc noop(NoopRequest) returns(NoopResponse);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/TestProcedure.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/TestProcedure.proto b/hbase-server/src/test/protobuf/TestProcedure.proto
deleted file mode 100644
index de74f36..0000000
--- a/hbase-server/src/test/protobuf/TestProcedure.proto
+++ /dev/null
@@ -1,24 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestProcedureProtos";
-option java_generic_services = true;
-
-message TestTableDDLStateData {
-  required string table_name = 1;
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/test.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/test.proto b/hbase-server/src/test/protobuf/test.proto
deleted file mode 100644
index 566b04b..0000000
--- a/hbase-server/src/test/protobuf/test.proto
+++ /dev/null
@@ -1,35 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestProtos";
-option java_generate_equals_and_hash = true;
-
-message EmptyRequestProto {
-}
-
-message EmptyResponseProto {
-}
-
-message EchoRequestProto {
-  required string message = 1;
-}
-
-message EchoResponseProto {
-  required string message = 1;
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-server/src/test/protobuf/test_rpc_service.proto
----------------------------------------------------------------------
diff --git a/hbase-server/src/test/protobuf/test_rpc_service.proto b/hbase-server/src/test/protobuf/test_rpc_service.proto
deleted file mode 100644
index 4ed0380..0000000
--- a/hbase-server/src/test/protobuf/test_rpc_service.proto
+++ /dev/null
@@ -1,33 +0,0 @@
-/**
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements.  See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership.  The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License.  You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-option java_package = "org.apache.hadoop.hbase.ipc.protobuf.generated";
-option java_outer_classname = "TestRpcServiceProtos";
-option java_generic_services = true;
-option java_generate_equals_and_hash = true;
-
-import "test.proto";
-
-
-/**
- * A protobuf service for use in tests
- */
-service TestProtobufRpcProto {
-  rpc ping(EmptyRequestProto) returns (EmptyResponseProto);
-  rpc echo(EchoRequestProto) returns (EchoResponseProto);
-  rpc error(EmptyRequestProto) returns (EmptyResponseProto);
-}

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-spark/pom.xml
----------------------------------------------------------------------
diff --git a/hbase-spark/pom.xml b/hbase-spark/pom.xml
index 776eadb..89ed186 100644
--- a/hbase-spark/pom.xml
+++ b/hbase-spark/pom.xml
@@ -663,43 +663,5 @@
                 <surefire.skipSecondPart>true</surefire.skipSecondPart>
             </properties>
         </profile>
-        <profile>
-            <id>compile-protobuf</id>
-            <activation>
-                <property>
-                    <name>compile-protobuf</name>
-                </property>
-            </activation>
-            <build>
-                <plugins>
-                    <plugin>
-                        <groupId>org.apache.hadoop</groupId>
-                        <artifactId>hadoop-maven-plugins</artifactId>
-                        <executions>
-                            <execution>
-                                <id>compile-protoc</id>
-                                <phase>generate-sources</phase>
-                                <goals>
-                                    <goal>protoc</goal>
-                                </goals>
-                                <configuration>
-                                    <imports>
-                                        <param>${basedir}/src/main/protobuf</param>
-                                    </imports>
-                                    <source>
-                                        <directory>${basedir}/src/main/protobuf</directory>
-                                        <includes>
-                                            <include>Filter.proto</include>
-                                        </includes>
-                                    </source>
-                                    <!--<output>${project.build.directory}/generated-sources/java</output>-->
-                                    <output>${basedir}/src/main/java/</output>
-                                </configuration>
-                            </execution>
-                        </executions>
-                    </plugin>
-                </plugins>
-            </build>
-        </profile>
     </profiles>
 </project>

http://git-wip-us.apache.org/repos/asf/hbase/blob/9d740f7b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
----------------------------------------------------------------------
diff --git a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
index 071c1ca..fa66d69 100644
--- a/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
+++ b/hbase-spark/src/main/java/org/apache/hadoop/hbase/spark/SparkSQLPushDownFilter.java
@@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.exceptions.DeserializationException;
 import org.apache.hadoop.hbase.filter.FilterBase;
 import org.apache.hadoop.hbase.spark.datasources.BytesEncoder;
 import org.apache.hadoop.hbase.spark.datasources.JavaBytesEncoder;
-import org.apache.hadoop.hbase.spark.protobuf.generated.FilterProtos;
+import org.apache.hadoop.hbase.spark.protobuf.generated.SparkFilterProtos;
 import org.apache.hadoop.hbase.util.ByteStringer;
 import org.apache.hadoop.hbase.util.Bytes;
 import org.apache.spark.sql.datasources.hbase.Field;
@@ -183,9 +183,9 @@ public class SparkSQLPushDownFilter extends FilterBase{
   public static SparkSQLPushDownFilter parseFrom(final byte[] pbBytes)
           throws DeserializationException {
 
-    FilterProtos.SQLPredicatePushDownFilter proto;
+    SparkFilterProtos.SQLPredicatePushDownFilter proto;
     try {
-      proto = FilterProtos.SQLPredicatePushDownFilter.parseFrom(pbBytes);
+      proto = SparkFilterProtos.SQLPredicatePushDownFilter.parseFrom(pbBytes);
     } catch (InvalidProtocolBufferException e) {
       throw new DeserializationException(e);
     }
@@ -208,7 +208,7 @@ public class SparkSQLPushDownFilter extends FilterBase{
     HashMap<ByteArrayComparable, HashMap<ByteArrayComparable, String>>
             currentCellToColumnIndexMap = new HashMap<>();
 
-    for (FilterProtos.SQLPredicatePushDownCellToColumnMapping
+    for (SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping
             sqlPredicatePushDownCellToColumnMapping :
             proto.getCellToColumnMappingList()) {
 
@@ -242,11 +242,11 @@ public class SparkSQLPushDownFilter extends FilterBase{
    */
   public byte[] toByteArray() {
 
-    FilterProtos.SQLPredicatePushDownFilter.Builder builder =
-            FilterProtos.SQLPredicatePushDownFilter.newBuilder();
+    SparkFilterProtos.SQLPredicatePushDownFilter.Builder builder =
+            SparkFilterProtos.SQLPredicatePushDownFilter.newBuilder();
 
-    FilterProtos.SQLPredicatePushDownCellToColumnMapping.Builder columnMappingBuilder =
-            FilterProtos.SQLPredicatePushDownCellToColumnMapping.newBuilder();
+    SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping.Builder columnMappingBuilder
=
+            SparkFilterProtos.SQLPredicatePushDownCellToColumnMapping.newBuilder();
 
     builder.setDynamicLogicExpression(dynamicLogicExpression.toExpressionString());
     for (byte[] valueFromQuery: valueFromQueryArray) {


Mime
View raw message