hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject svn commit: r1522035 [8/8] - in /hbase/trunk: hbase-client/src/main/java/org/apache/hadoop/hbase/client/ hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/ hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ hbase-client/src/mai...
Date Wed, 11 Sep 2013 21:07:34 GMT
Added: hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TracingProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TracingProtos.java?rev=1522035&view=auto
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TracingProtos.java (added)
+++ hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/TracingProtos.java Wed Sep 11 21:07:33 2013
@@ -0,0 +1,591 @@
+// Generated by the protocol buffer compiler.  DO NOT EDIT!
+// source: Tracing.proto
+
+package org.apache.hadoop.hbase.protobuf.generated;
+
+public final class TracingProtos {
+  private TracingProtos() {}
+  public static void registerAllExtensions(
+      com.google.protobuf.ExtensionRegistry registry) {
+  }
+  public interface RPCTInfoOrBuilder
+      extends com.google.protobuf.MessageOrBuilder {
+
+    // optional int64 trace_id = 1;
+    /**
+     * <code>optional int64 trace_id = 1;</code>
+     */
+    boolean hasTraceId();
+    /**
+     * <code>optional int64 trace_id = 1;</code>
+     */
+    long getTraceId();
+
+    // optional int64 parent_id = 2;
+    /**
+     * <code>optional int64 parent_id = 2;</code>
+     */
+    boolean hasParentId();
+    /**
+     * <code>optional int64 parent_id = 2;</code>
+     */
+    long getParentId();
+  }
+  /**
+   * Protobuf type {@code RPCTInfo}
+   *
+   * <pre>
+   *Used to pass through the information necessary to continue
+   *a trace after an RPC is made. All we need is the traceid 
+   *(so we know the overarching trace this message is a part of), and
+   *the id of the current span when this message was sent, so we know 
+   *what span caused the new span we will create when this message is received.
+   * </pre>
+   */
+  public static final class RPCTInfo extends
+      com.google.protobuf.GeneratedMessage
+      implements RPCTInfoOrBuilder {
+    // Use RPCTInfo.newBuilder() to construct.
+    private RPCTInfo(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+      super(builder);
+      this.unknownFields = builder.getUnknownFields();
+    }
+    private RPCTInfo(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+
+    private static final RPCTInfo defaultInstance;
+    public static RPCTInfo getDefaultInstance() {
+      return defaultInstance;
+    }
+
+    public RPCTInfo getDefaultInstanceForType() {
+      return defaultInstance;
+    }
+
+    private final com.google.protobuf.UnknownFieldSet unknownFields;
+    @java.lang.Override
+    public final com.google.protobuf.UnknownFieldSet
+        getUnknownFields() {
+      return this.unknownFields;
+    }
+    private RPCTInfo(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      initFields();
+      int mutable_bitField0_ = 0;
+      com.google.protobuf.UnknownFieldSet.Builder unknownFields =
+          com.google.protobuf.UnknownFieldSet.newBuilder();
+      try {
+        boolean done = false;
+        while (!done) {
+          int tag = input.readTag();
+          switch (tag) {
+            case 0:
+              done = true;
+              break;
+            default: {
+              if (!parseUnknownField(input, unknownFields,
+                                     extensionRegistry, tag)) {
+                done = true;
+              }
+              break;
+            }
+            case 8: {
+              bitField0_ |= 0x00000001;
+              traceId_ = input.readInt64();
+              break;
+            }
+            case 16: {
+              bitField0_ |= 0x00000002;
+              parentId_ = input.readInt64();
+              break;
+            }
+          }
+        }
+      } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+        throw e.setUnfinishedMessage(this);
+      } catch (java.io.IOException e) {
+        throw new com.google.protobuf.InvalidProtocolBufferException(
+            e.getMessage()).setUnfinishedMessage(this);
+      } finally {
+        this.unknownFields = unknownFields.build();
+        makeExtensionsImmutable();
+      }
+    }
+    public static final com.google.protobuf.Descriptors.Descriptor
+        getDescriptor() {
+      return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.internal_static_RPCTInfo_descriptor;
+    }
+
+    protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+        internalGetFieldAccessorTable() {
+      return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.internal_static_RPCTInfo_fieldAccessorTable
+          .ensureFieldAccessorsInitialized(
+              org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.class, org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.Builder.class);
+    }
+
+    public static com.google.protobuf.Parser<RPCTInfo> PARSER =
+        new com.google.protobuf.AbstractParser<RPCTInfo>() {
+      public RPCTInfo parsePartialFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws com.google.protobuf.InvalidProtocolBufferException {
+        return new RPCTInfo(input, extensionRegistry);
+      }
+    };
+
+    @java.lang.Override
+    public com.google.protobuf.Parser<RPCTInfo> getParserForType() {
+      return PARSER;
+    }
+
+    private int bitField0_;
+    // optional int64 trace_id = 1;
+    public static final int TRACE_ID_FIELD_NUMBER = 1;
+    private long traceId_;
+    /**
+     * <code>optional int64 trace_id = 1;</code>
+     */
+    public boolean hasTraceId() {
+      return ((bitField0_ & 0x00000001) == 0x00000001);
+    }
+    /**
+     * <code>optional int64 trace_id = 1;</code>
+     */
+    public long getTraceId() {
+      return traceId_;
+    }
+
+    // optional int64 parent_id = 2;
+    public static final int PARENT_ID_FIELD_NUMBER = 2;
+    private long parentId_;
+    /**
+     * <code>optional int64 parent_id = 2;</code>
+     */
+    public boolean hasParentId() {
+      return ((bitField0_ & 0x00000002) == 0x00000002);
+    }
+    /**
+     * <code>optional int64 parent_id = 2;</code>
+     */
+    public long getParentId() {
+      return parentId_;
+    }
+
+    private void initFields() {
+      traceId_ = 0L;
+      parentId_ = 0L;
+    }
+    private byte memoizedIsInitialized = -1;
+    public final boolean isInitialized() {
+      byte isInitialized = memoizedIsInitialized;
+      if (isInitialized != -1) return isInitialized == 1;
+
+      memoizedIsInitialized = 1;
+      return true;
+    }
+
+    public void writeTo(com.google.protobuf.CodedOutputStream output)
+                        throws java.io.IOException {
+      getSerializedSize();
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        output.writeInt64(1, traceId_);
+      }
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+        output.writeInt64(2, parentId_);
+      }
+      getUnknownFields().writeTo(output);
+    }
+
+    private int memoizedSerializedSize = -1;
+    public int getSerializedSize() {
+      int size = memoizedSerializedSize;
+      if (size != -1) return size;
+
+      size = 0;
+      if (((bitField0_ & 0x00000001) == 0x00000001)) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeInt64Size(1, traceId_);
+      }
+      if (((bitField0_ & 0x00000002) == 0x00000002)) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeInt64Size(2, parentId_);
+      }
+      size += getUnknownFields().getSerializedSize();
+      memoizedSerializedSize = size;
+      return size;
+    }
+
+    private static final long serialVersionUID = 0L;
+    @java.lang.Override
+    protected java.lang.Object writeReplace()
+        throws java.io.ObjectStreamException {
+      return super.writeReplace();
+    }
+
+    @java.lang.Override
+    public boolean equals(final java.lang.Object obj) {
+      if (obj == this) {
+       return true;
+      }
+      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo)) {
+        return super.equals(obj);
+      }
+      org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo other = (org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo) obj;
+
+      boolean result = true;
+      result = result && (hasTraceId() == other.hasTraceId());
+      if (hasTraceId()) {
+        result = result && (getTraceId()
+            == other.getTraceId());
+      }
+      result = result && (hasParentId() == other.hasParentId());
+      if (hasParentId()) {
+        result = result && (getParentId()
+            == other.getParentId());
+      }
+      result = result &&
+          getUnknownFields().equals(other.getUnknownFields());
+      return result;
+    }
+
+    private int memoizedHashCode = 0;
+    @java.lang.Override
+    public int hashCode() {
+      if (memoizedHashCode != 0) {
+        return memoizedHashCode;
+      }
+      int hash = 41;
+      hash = (19 * hash) + getDescriptorForType().hashCode();
+      if (hasTraceId()) {
+        hash = (37 * hash) + TRACE_ID_FIELD_NUMBER;
+        hash = (53 * hash) + hashLong(getTraceId());
+      }
+      if (hasParentId()) {
+        hash = (37 * hash) + PARENT_ID_FIELD_NUMBER;
+        hash = (53 * hash) + hashLong(getParentId());
+      }
+      hash = (29 * hash) + getUnknownFields().hashCode();
+      memoizedHashCode = hash;
+      return hash;
+    }
+
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        com.google.protobuf.ByteString data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        com.google.protobuf.ByteString data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(byte[] data)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        byte[] data,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws com.google.protobuf.InvalidProtocolBufferException {
+      return PARSER.parseFrom(data, extensionRegistry);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(java.io.InputStream input)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input, extensionRegistry);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseDelimitedFrom(java.io.InputStream input)
+        throws java.io.IOException {
+      return PARSER.parseDelimitedFrom(input);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseDelimitedFrom(
+        java.io.InputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws java.io.IOException {
+      return PARSER.parseDelimitedFrom(input, extensionRegistry);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        com.google.protobuf.CodedInputStream input)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input);
+    }
+    public static org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parseFrom(
+        com.google.protobuf.CodedInputStream input,
+        com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+        throws java.io.IOException {
+      return PARSER.parseFrom(input, extensionRegistry);
+    }
+
+    public static Builder newBuilder() { return Builder.create(); }
+    public Builder newBuilderForType() { return newBuilder(); }
+    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo prototype) {
+      return newBuilder().mergeFrom(prototype);
+    }
+    public Builder toBuilder() { return newBuilder(this); }
+
+    @java.lang.Override
+    protected Builder newBuilderForType(
+        com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+      Builder builder = new Builder(parent);
+      return builder;
+    }
+    /**
+     * Protobuf type {@code RPCTInfo}
+     *
+     * <pre>
+     *Used to pass through the information necessary to continue
+     *a trace after an RPC is made. All we need is the traceid 
+     *(so we know the overarching trace this message is a part of), and
+     *the id of the current span when this message was sent, so we know 
+     *what span caused the new span we will create when this message is received.
+     * </pre>
+     */
+    public static final class Builder extends
+        com.google.protobuf.GeneratedMessage.Builder<Builder>
+       implements org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfoOrBuilder {
+      public static final com.google.protobuf.Descriptors.Descriptor
+          getDescriptor() {
+        return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.internal_static_RPCTInfo_descriptor;
+      }
+
+      protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
+          internalGetFieldAccessorTable() {
+        return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.internal_static_RPCTInfo_fieldAccessorTable
+            .ensureFieldAccessorsInitialized(
+                org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.class, org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.Builder.class);
+      }
+
+      // Construct using org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.newBuilder()
+      private Builder() {
+        maybeForceBuilderInitialization();
+      }
+
+      private Builder(
+          com.google.protobuf.GeneratedMessage.BuilderParent parent) {
+        super(parent);
+        maybeForceBuilderInitialization();
+      }
+      private void maybeForceBuilderInitialization() {
+        if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+        }
+      }
+      private static Builder create() {
+        return new Builder();
+      }
+
+      public Builder clear() {
+        super.clear();
+        traceId_ = 0L;
+        bitField0_ = (bitField0_ & ~0x00000001);
+        parentId_ = 0L;
+        bitField0_ = (bitField0_ & ~0x00000002);
+        return this;
+      }
+
+      public Builder clone() {
+        return create().mergeFrom(buildPartial());
+      }
+
+      public com.google.protobuf.Descriptors.Descriptor
+          getDescriptorForType() {
+        return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.internal_static_RPCTInfo_descriptor;
+      }
+
+      public org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo getDefaultInstanceForType() {
+        return org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.getDefaultInstance();
+      }
+
+      public org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo build() {
+        org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo result = buildPartial();
+        if (!result.isInitialized()) {
+          throw newUninitializedMessageException(result);
+        }
+        return result;
+      }
+
+      public org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo buildPartial() {
+        org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo result = new org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo(this);
+        int from_bitField0_ = bitField0_;
+        int to_bitField0_ = 0;
+        if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
+          to_bitField0_ |= 0x00000001;
+        }
+        result.traceId_ = traceId_;
+        if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
+          to_bitField0_ |= 0x00000002;
+        }
+        result.parentId_ = parentId_;
+        result.bitField0_ = to_bitField0_;
+        onBuilt();
+        return result;
+      }
+
+      public Builder mergeFrom(com.google.protobuf.Message other) {
+        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo) {
+          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo)other);
+        } else {
+          super.mergeFrom(other);
+          return this;
+        }
+      }
+
+      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo other) {
+        if (other == org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo.getDefaultInstance()) return this;
+        if (other.hasTraceId()) {
+          setTraceId(other.getTraceId());
+        }
+        if (other.hasParentId()) {
+          setParentId(other.getParentId());
+        }
+        this.mergeUnknownFields(other.getUnknownFields());
+        return this;
+      }
+
+      public final boolean isInitialized() {
+        return true;
+      }
+
+      public Builder mergeFrom(
+          com.google.protobuf.CodedInputStream input,
+          com.google.protobuf.ExtensionRegistryLite extensionRegistry)
+          throws java.io.IOException {
+        org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo parsedMessage = null;
+        try {
+          parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
+        } catch (com.google.protobuf.InvalidProtocolBufferException e) {
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.TracingProtos.RPCTInfo) e.getUnfinishedMessage();
+          throw e;
+        } finally {
+          if (parsedMessage != null) {
+            mergeFrom(parsedMessage);
+          }
+        }
+        return this;
+      }
+      private int bitField0_;
+
+      // optional int64 trace_id = 1;
+      private long traceId_ ;
+      /**
+       * <code>optional int64 trace_id = 1;</code>
+       */
+      public boolean hasTraceId() {
+        return ((bitField0_ & 0x00000001) == 0x00000001);
+      }
+      /**
+       * <code>optional int64 trace_id = 1;</code>
+       */
+      public long getTraceId() {
+        return traceId_;
+      }
+      /**
+       * <code>optional int64 trace_id = 1;</code>
+       */
+      public Builder setTraceId(long value) {
+        bitField0_ |= 0x00000001;
+        traceId_ = value;
+        onChanged();
+        return this;
+      }
+      /**
+       * <code>optional int64 trace_id = 1;</code>
+       */
+      public Builder clearTraceId() {
+        bitField0_ = (bitField0_ & ~0x00000001);
+        traceId_ = 0L;
+        onChanged();
+        return this;
+      }
+
+      // optional int64 parent_id = 2;
+      private long parentId_ ;
+      /**
+       * <code>optional int64 parent_id = 2;</code>
+       */
+      public boolean hasParentId() {
+        return ((bitField0_ & 0x00000002) == 0x00000002);
+      }
+      /**
+       * <code>optional int64 parent_id = 2;</code>
+       */
+      public long getParentId() {
+        return parentId_;
+      }
+      /**
+       * <code>optional int64 parent_id = 2;</code>
+       */
+      public Builder setParentId(long value) {
+        bitField0_ |= 0x00000002;
+        parentId_ = value;
+        onChanged();
+        return this;
+      }
+      /**
+       * <code>optional int64 parent_id = 2;</code>
+       */
+      public Builder clearParentId() {
+        bitField0_ = (bitField0_ & ~0x00000002);
+        parentId_ = 0L;
+        onChanged();
+        return this;
+      }
+
+      // @@protoc_insertion_point(builder_scope:RPCTInfo)
+    }
+
+    static {
+      defaultInstance = new RPCTInfo(true);
+      defaultInstance.initFields();
+    }
+
+    // @@protoc_insertion_point(class_scope:RPCTInfo)
+  }
+
+  private static com.google.protobuf.Descriptors.Descriptor
+    internal_static_RPCTInfo_descriptor;
+  private static
+    com.google.protobuf.GeneratedMessage.FieldAccessorTable
+      internal_static_RPCTInfo_fieldAccessorTable;
+
+  public static com.google.protobuf.Descriptors.FileDescriptor
+      getDescriptor() {
+    return descriptor;
+  }
+  private static com.google.protobuf.Descriptors.FileDescriptor
+      descriptor;
+  static {
+    java.lang.String[] descriptorData = {
+      "\n\rTracing.proto\"/\n\010RPCTInfo\022\020\n\010trace_id\030" +
+      "\001 \001(\003\022\021\n\tparent_id\030\002 \001(\003B@\n*org.apache.h" +
+      "adoop.hbase.protobuf.generatedB\rTracingP" +
+      "rotosH\001\240\001\001"
+    };
+    com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner assigner =
+      new com.google.protobuf.Descriptors.FileDescriptor.InternalDescriptorAssigner() {
+        public com.google.protobuf.ExtensionRegistry assignDescriptors(
+            com.google.protobuf.Descriptors.FileDescriptor root) {
+          descriptor = root;
+          internal_static_RPCTInfo_descriptor =
+            getDescriptor().getMessageTypes().get(0);
+          internal_static_RPCTInfo_fieldAccessorTable = new
+            com.google.protobuf.GeneratedMessage.FieldAccessorTable(
+              internal_static_RPCTInfo_descriptor,
+              new java.lang.String[] { "TraceId", "ParentId", });
+          return null;
+        }
+      };
+    com.google.protobuf.Descriptors.FileDescriptor
+      .internalBuildGeneratedFileFrom(descriptorData,
+        new com.google.protobuf.Descriptors.FileDescriptor[] {
+        }, assigner);
+  }
+
+  // @@protoc_insertion_point(outer_class_scope)
+}

Modified: hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java (original)
+++ hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/WALProtos.java Wed Sep 11 21:07:33 2013
@@ -4811,7 +4811,7 @@ public final class WALProtos {
       descriptor;
   static {
     java.lang.String[] descriptorData = {
-      "\n\tWAL.proto\032\013hbase.proto\"$\n\tWALHeader\022\027\n" +
+      "\n\tWAL.proto\032\013HBase.proto\"$\n\tWALHeader\022\027\n" +
       "\017has_compression\030\001 \001(\010\"\337\001\n\006WALKey\022\033\n\023enc" +
       "oded_region_name\030\001 \002(\014\022\022\n\ntable_name\030\002 \002" +
       "(\014\022\033\n\023log_sequence_number\030\003 \002(\004\022\022\n\nwrite" +

Modified: hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java (original)
+++ hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ZooKeeperProtos.java Wed Sep 11 21:07:33 2013
@@ -9391,7 +9391,7 @@ public final class ZooKeeperProtos {
       descriptor;
   static {
     java.lang.String[] descriptorData = {
-      "\n\017ZooKeeper.proto\032\013hbase.proto\"D\n\020MetaRe" +
+      "\n\017ZooKeeper.proto\032\013HBase.proto\"D\n\020MetaRe" +
       "gionServer\022\033\n\006server\030\001 \002(\0132\013.ServerName\022" +
       "\023\n\013rpc_version\030\002 \001(\r\":\n\006Master\022\033\n\006master" +
       "\030\001 \002(\0132\013.ServerName\022\023\n\013rpc_version\030\002 \001(\r" +

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/AccessControl.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/AccessControl.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/AccessControl.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/AccessControl.proto Wed Sep 11 21:07:33 2013
@@ -22,7 +22,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 
 message Permission {
     enum Action {
@@ -90,13 +90,13 @@ message RevokeRequest {
 message RevokeResponse {
 }
 
-message UserPermissionsRequest {
+message GetUserPermissionsRequest {
   optional Permission.Type type = 1;
   optional TableName table_name = 2;
   optional bytes namespace_name = 3;
 }
 
-message UserPermissionsResponse {
+message GetUserPermissionsResponse {
   repeated UserPermission user_permission = 1;
 }
 
@@ -114,8 +114,8 @@ service AccessControlService {
     rpc Revoke(RevokeRequest)
       returns (RevokeResponse);
 
-    rpc GetUserPermissions(UserPermissionsRequest)
-      returns (UserPermissionsResponse);
+    rpc GetUserPermissions(GetUserPermissionsRequest)
+      returns (GetUserPermissionsResponse);
 
     rpc CheckPermissions(CheckPermissionsRequest)
       returns (CheckPermissionsResponse);

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Admin.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Admin.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Admin.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Admin.proto Wed Sep 11 21:07:33 2013
@@ -25,7 +25,7 @@ option java_generate_equals_and_hash = t
 option optimize_for = SPEED;
 
 import "Client.proto";
-import "hbase.proto";
+import "HBase.proto";
 import "WAL.proto";
 
 message GetRegionInfoRequest {

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Aggregate.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Aggregate.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Aggregate.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Aggregate.proto Wed Sep 11 21:07:33 2013
@@ -24,8 +24,8 @@ option optimize_for = SPEED;
 
 import "Client.proto";
 
-message AggregateArgument {
-  /** The argument passed to the AggregateService consists of three parts
+message AggregateRequest {
+  /** The request passed to the AggregateService consists of three parts
    *  (1) the (canonical) classname of the ColumnInterpreter implementation
    *  (2) the Scan query
    *  (3) any bytes required to construct the ColumnInterpreter object
@@ -52,11 +52,11 @@ message AggregateResponse {
  *  AggregateService method implementations and their functionality.
  */
 service AggregateService {
-  rpc GetMax (AggregateArgument) returns (AggregateResponse);
-  rpc GetMin (AggregateArgument) returns (AggregateResponse);
-  rpc GetSum (AggregateArgument) returns (AggregateResponse);
-  rpc GetRowNum (AggregateArgument) returns (AggregateResponse);
-  rpc GetAvg (AggregateArgument) returns (AggregateResponse);
-  rpc GetStd (AggregateArgument) returns (AggregateResponse);
-  rpc GetMedian (AggregateArgument) returns (AggregateResponse);
+  rpc GetMax (AggregateRequest) returns (AggregateResponse);
+  rpc GetMin (AggregateRequest) returns (AggregateResponse);
+  rpc GetSum (AggregateRequest) returns (AggregateResponse);
+  rpc GetRowNum (AggregateRequest) returns (AggregateResponse);
+  rpc GetAvg (AggregateRequest) returns (AggregateResponse);
+  rpc GetStd (AggregateRequest) returns (AggregateResponse);
+  rpc GetMedian (AggregateRequest) returns (AggregateResponse);
 }

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Authentication.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Authentication.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Authentication.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Authentication.proto Wed Sep 11 21:07:33 2013
@@ -55,10 +55,10 @@ message Token {
 
 
 // RPC request & response messages
-message TokenRequest {
+message GetAuthenticationTokenRequest {
 }
 
-message TokenResponse {
+message GetAuthenticationTokenResponse {
     optional Token token = 1;
 }
 
@@ -73,8 +73,8 @@ message WhoAmIResponse {
 
 // RPC service
 service AuthenticationService {
-    rpc GetAuthenticationToken(TokenRequest)
-        returns (TokenResponse);
+    rpc GetAuthenticationToken(GetAuthenticationTokenRequest)
+        returns (GetAuthenticationTokenResponse);
 
     rpc WhoAmI(WhoAmIRequest)
         returns (WhoAmIResponse);

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Client.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Client.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Client.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Client.proto Wed Sep 11 21:07:33 2013
@@ -24,7 +24,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "Filter.proto";
 import "Cell.proto";
 import "Comparator.proto";

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/ClusterStatus.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/ClusterStatus.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/ClusterStatus.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/ClusterStatus.proto Wed Sep 11 21:07:33 2013
@@ -23,7 +23,7 @@ option java_outer_classname = "ClusterSt
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "ClusterId.proto";
 import "FS.proto";
 

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Filter.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Filter.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Filter.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Filter.proto Wed Sep 11 21:07:33 2013
@@ -24,7 +24,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "Comparator.proto";
 
 message Filter {

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/HFile.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/HFile.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/HFile.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/HFile.proto Wed Sep 11 21:07:33 2013
@@ -21,7 +21,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 
 // Map of name/values
 message FileInfoProto {

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/MapReduce.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/MapReduce.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/MapReduce.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/MapReduce.proto Wed Sep 11 21:07:33 2013
@@ -23,7 +23,7 @@
  option java_generate_equals_and_hash = true;
  option optimize_for = SPEED;
 
- import "hbase.proto";
+ import "HBase.proto";
 
  message ScanMetrics {
 

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/MasterAdmin.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/MasterAdmin.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/MasterAdmin.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/MasterAdmin.proto Wed Sep 11 21:07:33 2013
@@ -26,7 +26,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "Client.proto";
 
 /* Column-level protobufs */
@@ -158,6 +158,9 @@ message ModifyNamespaceRequest {
   required NamespaceDescriptor namespaceDescriptor = 1;
 }
 
+message ModifyNamespaceResponse {
+}
+
 message GetNamespaceDescriptorRequest {
   required string namespaceName = 1;
 }
@@ -166,9 +169,6 @@ message GetNamespaceDescriptorResponse {
   required NamespaceDescriptor namespaceDescriptor = 1;
 }
 
-message ModifyNamespaceResponse {
-}
-
 message ListNamespaceDescriptorsRequest {
 }
 
@@ -223,10 +223,10 @@ message SetBalancerRunningResponse {
   optional bool prev_balance_value = 1;
 }
 
-message CatalogScanRequest {
+message RunCatalogScanRequest {
 }
 
-message CatalogScanResponse {
+message RunCatalogScanResponse {
   optional int32 scan_result = 1;
 }
 
@@ -245,26 +245,26 @@ message IsCatalogJanitorEnabledResponse 
   required bool value = 1;
 }
 
-message TakeSnapshotRequest{
+message SnapshotRequest {
 	required SnapshotDescription snapshot = 1;
 }
 
-message TakeSnapshotResponse{
+message SnapshotResponse {
 	required int64 expected_timeout = 1;
 }
 
-message ListSnapshotRequest{
+message GetCompletedSnapshotsRequest {
 }
 
-message ListSnapshotResponse{
+message GetCompletedSnapshotsResponse {
 	repeated SnapshotDescription snapshots = 1;
 }
 
-message DeleteSnapshotRequest{
+message DeleteSnapshotRequest {
 	required SnapshotDescription snapshot = 1;
 }
 
-message DeleteSnapshotResponse{
+message DeleteSnapshotResponse {
 }
 
 message RestoreSnapshotRequest {
@@ -277,11 +277,11 @@ message RestoreSnapshotResponse {
 /* if you don't send the snapshot, then you will get it back
  * in the response (if the snapshot is done) so you can check the snapshot
  */
-message IsSnapshotDoneRequest{
+message IsSnapshotDoneRequest {
 	optional SnapshotDescription snapshot = 1;
 }
 
-message IsSnapshotDoneResponse{
+message IsSnapshotDoneResponse {
 	optional bool done = 1 [default = false];
 	optional SnapshotDescription snapshot = 2;
 }
@@ -381,8 +381,8 @@ service MasterAdminService {
     returns(SetBalancerRunningResponse);
 
   /** Get a run of the catalog janitor */
-  rpc RunCatalogScan(CatalogScanRequest)
-     returns(CatalogScanResponse);
+  rpc RunCatalogScan(RunCatalogScanRequest)
+     returns(RunCatalogScanResponse);
 
   /**
    * Enable the catalog janitor on or off.
@@ -405,13 +405,13 @@ service MasterAdminService {
   /** 
    * Create a snapshot for the given table.
    */
-  rpc Snapshot(TakeSnapshotRequest) returns(TakeSnapshotResponse);
+  rpc Snapshot(SnapshotRequest) returns(SnapshotResponse);
 
   /**
-   * List completed snapshots.
+   * Get completed snapshots.
    * Returns a list of snapshot descriptors for completed snapshots
    */
-  rpc GetCompletedSnapshots(ListSnapshotRequest) returns(ListSnapshotResponse);
+  rpc GetCompletedSnapshots(GetCompletedSnapshotsRequest) returns(GetCompletedSnapshotsResponse);
 
   /**
    * Delete an existing snapshot. This method can also be used to clean up an aborted snapshot.
@@ -445,7 +445,7 @@ service MasterAdminService {
   rpc CreateNamespace(CreateNamespaceRequest)
     returns(CreateNamespaceResponse);
 
-  /** Delete's namespace synchronously */
+  /** Deletes namespace synchronously */
   rpc DeleteNamespace(DeleteNamespaceRequest)
     returns(DeleteNamespaceResponse);
 

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/MasterMonitor.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/MasterMonitor.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/MasterMonitor.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/MasterMonitor.proto Wed Sep 11 21:07:33 2013
@@ -25,7 +25,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "ClusterStatus.proto";
 
 message GetSchemaAlterStatusRequest {

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/MultiRowMutation.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/MultiRowMutation.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/MultiRowMutation.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/MultiRowMutation.proto Wed Sep 11 21:07:33 2013
@@ -17,19 +17,19 @@
  */
 import "Client.proto";
 option java_package = "org.apache.hadoop.hbase.protobuf.generated";
-option java_outer_classname = "MultiRowMutation";
+option java_outer_classname = "MultiRowMutationProtos";
 option java_generate_equals_and_hash = true;
 option java_generic_services = true;
 option optimize_for = SPEED;
 
-message MultiMutateRequest {
+message MutateRowsRequest {
   repeated MutationProto mutation_request = 1;
 }
 
-message MultiMutateResponse {
+message MutateRowsResponse {
 }
 
 service MultiRowMutationService {
-  rpc MutateRows(MultiMutateRequest)
-      returns(MultiMutateResponse);
-}
+  rpc MutateRows(MutateRowsRequest)
+      returns(MutateRowsResponse);
+}
\ No newline at end of file

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/RPC.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/RPC.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/RPC.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/RPC.proto Wed Sep 11 21:07:33 2013
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 import "Tracing.proto"; 
-import "hbase.proto"; 
+import "HBase.proto";
 
 option java_package = "org.apache.hadoop.hbase.protobuf.generated";
 option java_outer_classname = "RPCProtos";

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/RegionServerStatus.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/RegionServerStatus.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/RegionServerStatus.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/RegionServerStatus.proto Wed Sep 11 21:07:33 2013
@@ -24,7 +24,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 import "ClusterStatus.proto";
 
 message RegionServerStartupRequest {

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/RowProcessor.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/RowProcessor.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/RowProcessor.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/RowProcessor.proto Wed Sep 11 21:07:33 2013
@@ -26,16 +26,16 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-message RowProcessorRequest {
+message ProcessRequest {
   required string row_processor_class_name = 1;
   optional string row_processor_initializer_message_name = 2;
   optional bytes  row_processor_initializer_message = 3;
 }
 
-message RowProcessorResult {
+message ProcessResponse {
   required bytes row_processor_result = 1;
 }
 
 service RowProcessorService {
-  rpc Process (RowProcessorRequest) returns (RowProcessorResult);
+  rpc Process(ProcessRequest) returns (ProcessResponse);
 }

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/SecureBulkLoad.proto Wed Sep 11 21:07:33 2013
@@ -22,13 +22,13 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import 'hbase.proto';
+import 'HBase.proto';
 import 'Client.proto';
 
 message SecureBulkLoadHFilesRequest {
   repeated BulkLoadHFileRequest.FamilyPath family_path = 1;
   optional bool assign_seq_num = 2;
-  required DelegationTokenProto fs_token = 3;
+  required DelegationToken fs_token = 3;
   required string bulk_token = 4;
 }
 
@@ -36,7 +36,7 @@ message SecureBulkLoadHFilesResponse {
   required bool loaded = 1;
 }
 
-message DelegationTokenProto {
+message DelegationToken {
   optional bytes identifier = 1;
   optional bytes password = 2;
   optional string kind = 3;

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/Tracing.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/Tracing.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/Tracing.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/Tracing.proto Wed Sep 11 21:07:33 2013
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 option java_package = "org.apache.hadoop.hbase.protobuf.generated";
-option java_outer_classname = "Tracing";
+option java_outer_classname = "TracingProtos";
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/WAL.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/WAL.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/WAL.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/WAL.proto Wed Sep 11 21:07:33 2013
@@ -21,7 +21,7 @@ option java_generic_services = false;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 
 message WALHeader {
   optional bool has_compression = 1;

Modified: hbase/trunk/hbase-protocol/src/main/protobuf/ZooKeeper.proto
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/protobuf/ZooKeeper.proto?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/protobuf/ZooKeeper.proto (original)
+++ hbase/trunk/hbase-protocol/src/main/protobuf/ZooKeeper.proto Wed Sep 11 21:07:33 2013
@@ -25,7 +25,7 @@ option java_generic_services = true;
 option java_generate_equals_and_hash = true;
 option optimize_for = SPEED;
 
-import "hbase.proto";
+import "HBase.proto";
 
 /**
  * Content of the meta-region-server znode.

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/catalog/MetaEditor.java Wed Sep 11 21:07:33 2013
@@ -39,8 +39,8 @@ import org.apache.hadoop.hbase.client.Pu
 import org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel;
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiMutateRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiRowMutationService;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateRowsRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MultiRowMutationService;
 import org.apache.hadoop.hbase.util.Bytes;
 
 import com.google.protobuf.ServiceException;
@@ -378,7 +378,7 @@ public class MetaEditor {
    */
   private static void multiMutate(HTable table, byte[] row, Mutation... mutations) throws IOException {
     CoprocessorRpcChannel channel = table.coprocessorService(row);
-    MultiMutateRequest.Builder mmrBuilder = MultiMutateRequest.newBuilder();
+    MutateRowsRequest.Builder mmrBuilder = MutateRowsRequest.newBuilder();
     for (Mutation mutation : mutations) {
       if (mutation instanceof Put) {
         mmrBuilder.addMutationRequest(ProtobufUtil.toMutation(MutationType.PUT, mutation));

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/client/coprocessor/RowProcessorClient.java Wed Sep 11 21:07:33 2013
@@ -22,7 +22,7 @@ import java.io.IOException;
 
 import org.apache.hadoop.classification.InterfaceAudience;
 import org.apache.hadoop.classification.InterfaceStability;
-import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorRequest;
+import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessRequest;
 import org.apache.hadoop.hbase.regionserver.RowProcessor;
 
 import com.google.protobuf.Message;
@@ -35,10 +35,10 @@ import com.google.protobuf.Message;
 @InterfaceStability.Evolving
 public class RowProcessorClient {
   public static <S extends Message, T extends Message>
-  RowProcessorRequest getRowProcessorPB(RowProcessor<S,T> r)
+  ProcessRequest getRowProcessorPB(RowProcessor<S,T> r)
       throws IOException {
-    final RowProcessorRequest.Builder requestBuilder =
-        RowProcessorRequest.newBuilder();
+    final ProcessRequest.Builder requestBuilder =
+        ProcessRequest.newBuilder();
     requestBuilder.setRowProcessorClassName(r.getClass().getName());
     S s = r.getRequestData();
     if (s != null) {

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/AggregateImplementation.java Wed Sep 11 21:07:33 2013
@@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.client.Sc
 import org.apache.hadoop.hbase.filter.FirstKeyOnlyFilter;
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.protobuf.ResponseConverter;
-import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateArgument;
+import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateRequest;
 import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateResponse;
 import org.apache.hadoop.hbase.protobuf.generated.AggregateProtos.AggregateService;
 import org.apache.hadoop.hbase.regionserver.InternalScanner;
@@ -73,7 +73,7 @@ extends AggregateService implements Copr
    * entire column family will be returned.
    */
   @Override
-  public void getMax(RpcController controller, AggregateArgument request,
+  public void getMax(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     InternalScanner scanner = null;
     AggregateResponse response = null;
@@ -127,7 +127,7 @@ extends AggregateService implements Copr
    * entire column family will be returned.
    */
   @Override
-  public void getMin(RpcController controller, AggregateArgument request,
+  public void getMin(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     AggregateResponse response = null;
     InternalScanner scanner = null;
@@ -179,7 +179,7 @@ extends AggregateService implements Copr
    * family will be returned.
    */
   @Override
-  public void getSum(RpcController controller, AggregateArgument request,
+  public void getSum(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     AggregateResponse response = null;
     InternalScanner scanner = null;
@@ -231,7 +231,7 @@ extends AggregateService implements Copr
    * @throws IOException
    */
   @Override
-  public void getRowNum(RpcController controller, AggregateArgument request,
+  public void getRowNum(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     AggregateResponse response = null;
     long counter = 0l;
@@ -288,7 +288,7 @@ extends AggregateService implements Copr
    * type.
    */
   @Override
-  public void getAvg(RpcController controller, AggregateArgument request,
+  public void getAvg(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     AggregateResponse response = null;
     InternalScanner scanner = null;
@@ -347,7 +347,7 @@ extends AggregateService implements Copr
    * deviation is square root of variance.
    */
   @Override
-  public void getStd(RpcController controller, AggregateArgument request,
+  public void getStd(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     InternalScanner scanner = null;
     AggregateResponse response = null;
@@ -411,7 +411,7 @@ extends AggregateService implements Copr
    * the second qualifier (optional) is for weight column.
    */
   @Override
-  public void getMedian(RpcController controller, AggregateArgument request,
+  public void getMedian(RpcController controller, AggregateRequest request,
       RpcCallback<AggregateResponse> done) {
     AggregateResponse response = null;
     InternalScanner scanner = null;
@@ -469,7 +469,7 @@ extends AggregateService implements Copr
 
   @SuppressWarnings("unchecked")
   ColumnInterpreter<T,S,P,Q,R> constructColumnInterpreterFromRequest(
-      AggregateArgument request) throws IOException {
+      AggregateRequest request) throws IOException {
     String className = request.getInterpreterClassName();
     Class<?> cls;
     try {

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/BaseRowProcessorEndpoint.java Wed Sep 11 21:07:33 2013
@@ -26,8 +26,8 @@ import org.apache.hadoop.classification.
 import org.apache.hadoop.hbase.Coprocessor;
 import org.apache.hadoop.hbase.CoprocessorEnvironment;
 import org.apache.hadoop.hbase.protobuf.ResponseConverter;
-import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorRequest;
-import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorResult;
+import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessRequest;
+import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessResponse;
 import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorService;
 import org.apache.hadoop.hbase.regionserver.HRegion;
 import org.apache.hadoop.hbase.regionserver.RowProcessor;
@@ -61,16 +61,16 @@ extends RowProcessorService implements C
    * the read-modify-write procedure.
    */
   @Override
-  public void process(RpcController controller, RowProcessorRequest request,
-      RpcCallback<RowProcessorResult> done) {
-    RowProcessorResult resultProto = null;
+  public void process(RpcController controller, ProcessRequest request,
+      RpcCallback<ProcessResponse> done) {
+    ProcessResponse resultProto = null;
     try {
       RowProcessor<S,T> processor = constructRowProcessorFromRequest(request);
       HRegion region = env.getRegion();
       region.processRowsWithLocks(processor);
       T result = processor.getResult();
-      RowProcessorResult.Builder b = RowProcessorResult.newBuilder();
-      b.setRowProcessorResult(result.toByteString()); 
+      ProcessResponse.Builder b = ProcessResponse.newBuilder();
+      b.setRowProcessorResult(result.toByteString());
       resultProto = b.build();
     } catch (Exception e) {
       ResponseConverter.setControllerException(controller, new IOException(e));
@@ -108,7 +108,7 @@ extends RowProcessorService implements C
   }
 
   @SuppressWarnings("unchecked")
-  RowProcessor<S,T> constructRowProcessorFromRequest(RowProcessorRequest request)
+  RowProcessor<S,T> constructRowProcessorFromRequest(ProcessRequest request)
       throws IOException {
     String className = request.getRowProcessorClassName();
     Class<?> cls;

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/coprocessor/MultiRowMutationEndpoint.java Wed Sep 11 21:07:33 2013
@@ -35,9 +35,9 @@ import org.apache.hadoop.hbase.util.Byte
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.protobuf.ResponseConverter;
 import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiMutateRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiMutateResponse;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiRowMutationService;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateRowsRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateRowsResponse;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MultiRowMutationService;
 
 import com.google.protobuf.RpcCallback;
 import com.google.protobuf.RpcController;
@@ -63,13 +63,13 @@ import com.google.protobuf.Service;
  * ...
  * Mutate m1 = ProtobufUtil.toMutate(MutateType.PUT, p1);
  * Mutate m2 = ProtobufUtil.toMutate(MutateType.PUT, p2);
- * MultiMutateRequest.Builder mrmBuilder = MultiMutateRequest.newBuilder();
+ * MutateRowsRequest.Builder mrmBuilder = MutateRowsRequest.newBuilder();
  * mrmBuilder.addMutationRequest(m1);
  * mrmBuilder.addMutationRequest(m2);
  * CoprocessorRpcChannel channel = t.coprocessorService(ROW);
  * MultiRowMutationService.BlockingInterface service = 
  *    MultiRowMutationService.newBlockingStub(channel);
- * MultiMutateRequest mrm = mrmBuilder.build();
+ * MutateRowsRequest mrm = mrmBuilder.build();
  * service.mutateRows(null, mrm);
  * </pre></code>
  */
@@ -79,9 +79,9 @@ public class MultiRowMutationEndpoint ex
 CoprocessorService, Coprocessor {
   private RegionCoprocessorEnvironment env;
   @Override
-  public void mutateRows(RpcController controller, MultiMutateRequest request, 
-      RpcCallback<MultiMutateResponse> done) {
-    MultiMutateResponse response = MultiMutateResponse.getDefaultInstance();
+  public void mutateRows(RpcController controller, MutateRowsRequest request, 
+      RpcCallback<MutateRowsResponse> done) {
+    MutateRowsResponse response = MutateRowsResponse.getDefaultInstance();
     try {
       // set of rows to lock, sorted to avoid deadlocks
       SortedSet<byte[]> rowsToLock = new TreeSet<byte[]>(Bytes.BYTES_COMPARATOR);

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/master/HMaster.java Wed Sep 11 21:07:33 2013
@@ -124,8 +124,8 @@ import org.apache.hadoop.hbase.protobuf.
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.AssignRegionResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.BalanceRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.BalanceResponse;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.CatalogScanRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.CatalogScanResponse;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RunCatalogScanRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.RunCatalogScanResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.CreateTableRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.CreateTableResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.DeleteColumnRequest;
@@ -148,8 +148,8 @@ import org.apache.hadoop.hbase.protobuf.
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsRestoreSnapshotDoneResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsSnapshotDoneRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsSnapshotDoneResponse;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ListSnapshotRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ListSnapshotResponse;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.GetCompletedSnapshotsRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.GetCompletedSnapshotsResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ModifyColumnRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ModifyColumnResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ModifyTableRequest;
@@ -166,8 +166,8 @@ import org.apache.hadoop.hbase.protobuf.
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ShutdownResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.StopMasterRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.StopMasterResponse;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.TakeSnapshotRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.TakeSnapshotResponse;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.SnapshotRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.SnapshotResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.UnassignRegionRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.UnassignRegionResponse;
 import org.apache.hadoop.hbase.protobuf.generated.MasterMonitorProtos;
@@ -1459,10 +1459,10 @@ MasterServices, Server {
   }
 
   @Override
-  public CatalogScanResponse runCatalogScan(RpcController c,
-      CatalogScanRequest req) throws ServiceException {
+  public RunCatalogScanResponse runCatalogScan(RpcController c,
+      RunCatalogScanRequest req) throws ServiceException {
     try {
-      return ResponseConverter.buildCatalogScanResponse(catalogJanitorChore.scan());
+      return ResponseConverter.buildRunCatalogScanResponse(catalogJanitorChore.scan());
     } catch (IOException ioe) {
       throw new ServiceException(ioe);
     }
@@ -2875,7 +2875,7 @@ MasterServices, Server {
    * {@inheritDoc}
    */
   @Override
-  public TakeSnapshotResponse snapshot(RpcController controller, TakeSnapshotRequest request)
+  public SnapshotResponse snapshot(RpcController controller, SnapshotRequest request)
       throws ServiceException {
     try {
       this.snapshotManager.checkSnapshotSupport();
@@ -2897,17 +2897,17 @@ MasterServices, Server {
     // send back the max amount of time the client should wait for the snapshot to complete
     long waitTime = SnapshotDescriptionUtils.getMaxMasterTimeout(conf, snapshot.getType(),
       SnapshotDescriptionUtils.DEFAULT_MAX_WAIT_TIME);
-    return TakeSnapshotResponse.newBuilder().setExpectedTimeout(waitTime).build();
+    return SnapshotResponse.newBuilder().setExpectedTimeout(waitTime).build();
   }
 
   /**
    * List the currently available/stored snapshots. Any in-progress snapshots are ignored
    */
   @Override
-  public ListSnapshotResponse getCompletedSnapshots(RpcController controller,
-      ListSnapshotRequest request) throws ServiceException {
+  public GetCompletedSnapshotsResponse getCompletedSnapshots(RpcController controller,
+      GetCompletedSnapshotsRequest request) throws ServiceException {
     try {
-      ListSnapshotResponse.Builder builder = ListSnapshotResponse.newBuilder();
+      GetCompletedSnapshotsResponse.Builder builder = GetCompletedSnapshotsResponse.newBuilder();
       List<SnapshotDescription> snapshots = snapshotManager.getCompletedSnapshots();
 
       // convert to protobuf

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/access/AccessController.java Wed Sep 11 21:07:33 2013
@@ -1324,9 +1324,9 @@ public class AccessController extends Ba
 
   @Override
   public void getUserPermissions(RpcController controller,
-                                 AccessControlProtos.UserPermissionsRequest request,
-                                 RpcCallback<AccessControlProtos.UserPermissionsResponse> done) {
-    AccessControlProtos.UserPermissionsResponse response = null;
+                                 AccessControlProtos.GetUserPermissionsRequest request,
+                                 RpcCallback<AccessControlProtos.GetUserPermissionsResponse> done) {
+    AccessControlProtos.GetUserPermissionsResponse response = null;
     try {
       // only allowed to be called on _acl_ region
       if (aclRegion) {
@@ -1347,7 +1347,7 @@ public class AccessController extends Ba
           perms = AccessControlLists.getUserPermissions(
               regionEnv.getConfiguration(), null);
         }
-        response = ResponseConverter.buildUserPermissionsResponse(perms);
+        response = ResponseConverter.buildGetUserPermissionsResponse(perms);
       } else {
         throw new CoprocessorException(AccessController.class, "This method "
             + "can only execute at " + AccessControlLists.ACL_TABLE_NAME + " table.");

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenProvider.java Wed Sep 11 21:07:33 2013
@@ -98,10 +98,10 @@ public class TokenProvider implements Au
 
   @Override
   public void getAuthenticationToken(RpcController controller,
-                                     AuthenticationProtos.TokenRequest request,
-                                     RpcCallback<AuthenticationProtos.TokenResponse> done) {
-    AuthenticationProtos.TokenResponse.Builder response =
-        AuthenticationProtos.TokenResponse.newBuilder();
+                                     AuthenticationProtos.GetAuthenticationTokenRequest request,
+                                     RpcCallback<AuthenticationProtos.GetAuthenticationTokenResponse> done) {
+    AuthenticationProtos.GetAuthenticationTokenResponse.Builder response =
+        AuthenticationProtos.GetAuthenticationTokenResponse.newBuilder();
 
     try {
       if (secretManager == null) {

Modified: hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java (original)
+++ hbase/trunk/hbase-server/src/main/java/org/apache/hadoop/hbase/security/token/TokenUtil.java Wed Sep 11 21:07:33 2013
@@ -57,8 +57,8 @@ public class TokenUtil {
       CoprocessorRpcChannel rpcChannel = meta.coprocessorService(HConstants.EMPTY_START_ROW);
       AuthenticationProtos.AuthenticationService.BlockingInterface service =
           AuthenticationProtos.AuthenticationService.newBlockingStub(rpcChannel);
-      AuthenticationProtos.TokenResponse response = service.getAuthenticationToken(null,
-          AuthenticationProtos.TokenRequest.getDefaultInstance());
+      AuthenticationProtos.GetAuthenticationTokenResponse response = service.getAuthenticationToken(null,
+          AuthenticationProtos.GetAuthenticationTokenRequest.getDefaultInstance());
 
       return ProtobufUtil.toToken(response.getToken());
     } catch (ServiceException se) {

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestFromClientSide.java Wed Sep 11 21:07:33 2013
@@ -85,8 +85,8 @@ import org.apache.hadoop.hbase.ipc.RpcSe
 import org.apache.hadoop.hbase.protobuf.ProtobufUtil;
 import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto;
 import org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.MutationType;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiMutateRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutation.MultiRowMutationService;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MutateRowsRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MultiRowMutationProtos.MultiRowMutationService;
 import org.apache.hadoop.hbase.regionserver.HRegion;
 import org.apache.hadoop.hbase.regionserver.HRegionServer;
 import org.apache.hadoop.hbase.regionserver.NoSuchColumnFamilyException;
@@ -4314,10 +4314,10 @@ public class TestFromClientSide {
     p.add(FAMILY, QUALIFIER, VALUE);
     MutationProto m2 = ProtobufUtil.toMutation(MutationType.PUT, p);
 
-    MultiMutateRequest.Builder mrmBuilder = MultiMutateRequest.newBuilder();
+    MutateRowsRequest.Builder mrmBuilder = MutateRowsRequest.newBuilder();
     mrmBuilder.addMutationRequest(m1);
     mrmBuilder.addMutationRequest(m2);
-    MultiMutateRequest mrm = mrmBuilder.build();
+    MutateRowsRequest mrm = mrmBuilder.build();
     CoprocessorRpcChannel channel = t.coprocessorService(ROW);
     MultiRowMutationService.BlockingInterface service =
        MultiRowMutationService.newBlockingStub(channel);

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/coprocessor/TestRowProcessorEndpoint.java Wed Sep 11 21:07:33 2013
@@ -55,8 +55,8 @@ import org.apache.hadoop.hbase.coprocess
 import org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.RowSwapProcessorResponse;
 import org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorRequest;
 import org.apache.hadoop.hbase.coprocessor.protobuf.generated.IncrementCounterProcessorTestProtos.TimeoutProcessorResponse;
-import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorRequest;
-import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorResult;
+import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessRequest;
+import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.ProcessResponse;
 import org.apache.hadoop.hbase.protobuf.generated.RowProcessorProtos.RowProcessorService;
 import org.apache.hadoop.hbase.regionserver.BaseRowProcessor;
 import org.apache.hadoop.hbase.regionserver.HRegion;
@@ -76,7 +76,7 @@ import com.sun.org.apache.commons.loggin
 import com.sun.org.apache.commons.logging.LogFactory;
 
 /**
- * Verifies ProcessRowEndpoint works.
+ * Verifies ProcessEndpoint works.
  * The tested RowProcessor performs two scans and a read-modify-write.
  */
 @Category(MediumTests.class)
@@ -151,18 +151,18 @@ public class TestRowProcessorEndpoint {
   @Test
   public void testDoubleScan() throws Throwable {
     prepareTestData();
-    
+
     CoprocessorRpcChannel channel = table.coprocessorService(ROW);
     RowProcessorEndpoint.FriendsOfFriendsProcessor processor =
         new RowProcessorEndpoint.FriendsOfFriendsProcessor(ROW, A);
-    RowProcessorService.BlockingInterface service = 
+    RowProcessorService.BlockingInterface service =
         RowProcessorService.newBlockingStub(channel);
-    RowProcessorRequest request = RowProcessorClient.getRowProcessorPB(processor);
-    RowProcessorResult protoResult = service.process(null, request);
-    FriendsOfFriendsProcessorResponse response = 
+    ProcessRequest request = RowProcessorClient.getRowProcessorPB(processor);
+    ProcessResponse protoResult = service.process(null, request);
+    FriendsOfFriendsProcessorResponse response =
         FriendsOfFriendsProcessorResponse.parseFrom(protoResult.getRowProcessorResult());
     Set<String> result = new HashSet<String>();
-    result.addAll(response.getResultList()); 
+    result.addAll(response.getResultList());
     Set<String> expected =
       new HashSet<String>(Arrays.asList(new String[]{"d", "e", "f", "g"}));
     Get get = new Get(ROW);
@@ -198,10 +198,10 @@ public class TestRowProcessorEndpoint {
     CoprocessorRpcChannel channel = table.coprocessorService(ROW);
     RowProcessorEndpoint.IncrementCounterProcessor processor =
         new RowProcessorEndpoint.IncrementCounterProcessor(ROW);
-    RowProcessorService.BlockingInterface service = 
+    RowProcessorService.BlockingInterface service =
         RowProcessorService.newBlockingStub(channel);
-    RowProcessorRequest request = RowProcessorClient.getRowProcessorPB(processor);
-    RowProcessorResult protoResult = service.process(null, request);
+    ProcessRequest request = RowProcessorClient.getRowProcessorPB(processor);
+    ProcessResponse protoResult = service.process(null, request);
     IncCounterProcessorResponse response = IncCounterProcessorResponse
         .parseFrom(protoResult.getRowProcessorResult());
     Integer result = response.getResponse();
@@ -261,9 +261,9 @@ public class TestRowProcessorEndpoint {
     CoprocessorRpcChannel channel = table.coprocessorService(ROW);
     RowProcessorEndpoint.RowSwapProcessor processor =
         new RowProcessorEndpoint.RowSwapProcessor(ROW, ROW2);
-    RowProcessorService.BlockingInterface service = 
+    RowProcessorService.BlockingInterface service =
         RowProcessorService.newBlockingStub(channel);
-    RowProcessorRequest request = RowProcessorClient.getRowProcessorPB(processor);
+    ProcessRequest request = RowProcessorClient.getRowProcessorPB(processor);
     service.process(null, request);
   }
 
@@ -273,9 +273,9 @@ public class TestRowProcessorEndpoint {
     CoprocessorRpcChannel channel = table.coprocessorService(ROW);
     RowProcessorEndpoint.TimeoutProcessor processor =
         new RowProcessorEndpoint.TimeoutProcessor(ROW);
-    RowProcessorService.BlockingInterface service = 
+    RowProcessorService.BlockingInterface service =
         RowProcessorService.newBlockingStub(channel);
-    RowProcessorRequest request = RowProcessorClient.getRowProcessorPB(processor);
+    ProcessRequest request = RowProcessorClient.getRowProcessorPB(processor);
     boolean exceptionCaught = false;
     try {
       service.process(null, request);

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/master/cleaner/TestSnapshotFromMaster.java Wed Sep 11 21:07:33 2013
@@ -48,8 +48,8 @@ import org.apache.hadoop.hbase.protobuf.
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.DeleteSnapshotRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsSnapshotDoneRequest;
 import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.IsSnapshotDoneResponse;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ListSnapshotRequest;
-import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.ListSnapshotResponse;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.GetCompletedSnapshotsRequest;
+import org.apache.hadoop.hbase.protobuf.generated.MasterAdminProtos.GetCompletedSnapshotsResponse;
 import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy;
 import org.apache.hadoop.hbase.regionserver.HRegion;
 import org.apache.hadoop.hbase.snapshot.SnapshotDescriptionUtils;
@@ -216,8 +216,8 @@ public class TestSnapshotFromMaster {
   @Test(timeout = 300000)
   public void testGetCompletedSnapshots() throws Exception {
     // first check when there are no snapshots
-    ListSnapshotRequest request = ListSnapshotRequest.newBuilder().build();
-    ListSnapshotResponse response = master.getCompletedSnapshots(null, request);
+    GetCompletedSnapshotsRequest request = GetCompletedSnapshotsRequest.newBuilder().build();
+    GetCompletedSnapshotsResponse response = master.getCompletedSnapshots(null, request);
     assertEquals("Found unexpected number of snapshots", 0, response.getSnapshotsCount());
 
     // write one snapshot to the fs

Modified: hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java?rev=1522035&r1=1522034&r2=1522035&view=diff
==============================================================================
--- hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java (original)
+++ hbase/trunk/hbase-server/src/test/java/org/apache/hadoop/hbase/security/token/TestTokenAuthentication.java Wed Sep 11 21:07:33 2013
@@ -258,14 +258,14 @@ public class TestTokenAuthentication {
     }
 
     @Override
-    public AuthenticationProtos.TokenResponse getAuthenticationToken(
-        RpcController controller, AuthenticationProtos.TokenRequest request)
+    public AuthenticationProtos.GetAuthenticationTokenResponse getAuthenticationToken(
+        RpcController controller, AuthenticationProtos.GetAuthenticationTokenRequest request)
       throws ServiceException {
       LOG.debug("Authentication token request from "+RequestContext.getRequestUserName());
       // ignore passed in controller -- it's always null
       ServerRpcController serverController = new ServerRpcController();
-      BlockingRpcCallback<AuthenticationProtos.TokenResponse> callback =
-          new BlockingRpcCallback<AuthenticationProtos.TokenResponse>();
+      BlockingRpcCallback<AuthenticationProtos.GetAuthenticationTokenResponse> callback =
+          new BlockingRpcCallback<AuthenticationProtos.GetAuthenticationTokenResponse>();
       getAuthenticationToken(serverController, request, callback);
       try {
         serverController.checkFailed();



Mime
View raw message