hbase-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From st...@apache.org
Subject svn commit: r1527648 [2/3] - in /hbase/trunk: hbase-client/src/main/java/org/apache/hadoop/hbase/client/ hbase-client/src/main/java/org/apache/hadoop/hbase/ipc/ hbase-client/src/main/java/org/apache/hadoop/hbase/protobuf/ hbase-protocol/ hbase-protocol...
Date Mon, 30 Sep 2013 16:53:00 GMT
Modified: hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java
URL: http://svn.apache.org/viewvc/hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java?rev=1527648&r1=1527647&r2=1527648&view=diff
==============================================================================
--- hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java (original)
+++ hbase/trunk/hbase-protocol/src/main/java/org/apache/hadoop/hbase/protobuf/generated/ClientProtos.java Mon Sep 30 16:52:59 2013
@@ -23667,62 +23667,90 @@ public final class ClientProtos {
     // @@protoc_insertion_point(class_scope:CoprocessorServiceResponse)
   }
 
-  public interface MultiActionOrBuilder
+  public interface RegionMutationOrBuilder
       extends com.google.protobuf.MessageOrBuilder {
 
-    // optional .MutationProto mutation = 1;
+    // required .RegionSpecifier region = 1;
+    /**
+     * <code>required .RegionSpecifier region = 1;</code>
+     */
+    boolean hasRegion();
+    /**
+     * <code>required .RegionSpecifier region = 1;</code>
+     */
+    org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion();
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>required .RegionSpecifier region = 1;</code>
      */
-    boolean hasMutation();
+    org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder();
+
+    // optional bool atomic = 2;
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>optional bool atomic = 2;</code>
+     *
+     * <pre>
+     * When set, run mutations as atomic unit.
+     * </pre>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation();
+    boolean hasAtomic();
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>optional bool atomic = 2;</code>
+     *
+     * <pre>
+     * When set, run mutations as atomic unit.
+     * </pre>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder();
+    boolean getAtomic();
 
-    // optional .Get get = 2;
+    // repeated .MutationProto mutation = 3;
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
+     */
+    java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> 
+        getMutationList();
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
+     */
+    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation(int index);
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>repeated .MutationProto mutation = 3;</code>
      */
-    boolean hasGet();
+    int getMutationCount();
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>repeated .MutationProto mutation = 3;</code>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get getGet();
+    java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> 
+        getMutationOrBuilderList();
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>repeated .MutationProto mutation = 3;</code>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder getGetOrBuilder();
+    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder(
+        int index);
   }
   /**
-   * Protobuf type {@code MultiAction}
+   * Protobuf type {@code RegionMutation}
    *
    * <pre>
    **
-   * An action that is part of MultiRequest.
-   * This is a union type - exactly one of the fields will be set.
+   * Mutations to run against a Region.
    * </pre>
    */
-  public static final class MultiAction extends
+  public static final class RegionMutation extends
       com.google.protobuf.GeneratedMessage
-      implements MultiActionOrBuilder {
-    // Use MultiAction.newBuilder() to construct.
-    private MultiAction(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+      implements RegionMutationOrBuilder {
+    // Use RegionMutation.newBuilder() to construct.
+    private RegionMutation(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
       this.unknownFields = builder.getUnknownFields();
     }
-    private MultiAction(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+    private RegionMutation(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
 
-    private static final MultiAction defaultInstance;
-    public static MultiAction getDefaultInstance() {
+    private static final RegionMutation defaultInstance;
+    public static RegionMutation getDefaultInstance() {
       return defaultInstance;
     }
 
-    public MultiAction getDefaultInstanceForType() {
+    public RegionMutation getDefaultInstanceForType() {
       return defaultInstance;
     }
 
@@ -23732,7 +23760,7 @@ public final class ClientProtos {
         getUnknownFields() {
       return this.unknownFields;
     }
-    private MultiAction(
+    private RegionMutation(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -23756,29 +23784,29 @@ public final class ClientProtos {
               break;
             }
             case 10: {
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder subBuilder = null;
+              org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder subBuilder = null;
               if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = mutation_.toBuilder();
+                subBuilder = region_.toBuilder();
               }
-              mutation_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.PARSER, extensionRegistry);
+              region_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.PARSER, extensionRegistry);
               if (subBuilder != null) {
-                subBuilder.mergeFrom(mutation_);
-                mutation_ = subBuilder.buildPartial();
+                subBuilder.mergeFrom(region_);
+                region_ = subBuilder.buildPartial();
               }
               bitField0_ |= 0x00000001;
               break;
             }
-            case 18: {
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder subBuilder = null;
-              if (((bitField0_ & 0x00000002) == 0x00000002)) {
-                subBuilder = get_.toBuilder();
-              }
-              get_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.PARSER, extensionRegistry);
-              if (subBuilder != null) {
-                subBuilder.mergeFrom(get_);
-                get_ = subBuilder.buildPartial();
-              }
+            case 16: {
               bitField0_ |= 0x00000002;
+              atomic_ = input.readBool();
+              break;
+            }
+            case 26: {
+              if (!((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+                mutation_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>();
+                mutable_bitField0_ |= 0x00000004;
+              }
+              mutation_.add(input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.PARSER, extensionRegistry));
               break;
             }
           }
@@ -23789,99 +23817,143 @@ public final class ClientProtos {
         throw new com.google.protobuf.InvalidProtocolBufferException(
             e.getMessage()).setUnfinishedMessage(this);
       } finally {
+        if (((mutable_bitField0_ & 0x00000004) == 0x00000004)) {
+          mutation_ = java.util.Collections.unmodifiableList(mutation_);
+        }
         this.unknownFields = unknownFields.build();
         makeExtensionsImmutable();
       }
     }
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_MultiAction_descriptor;
+      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_RegionMutation_descriptor;
     }
 
     protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
         internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_MultiAction_fieldAccessorTable
+      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_RegionMutation_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.Builder.class);
+              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<MultiAction> PARSER =
-        new com.google.protobuf.AbstractParser<MultiAction>() {
-      public MultiAction parsePartialFrom(
+    public static com.google.protobuf.Parser<RegionMutation> PARSER =
+        new com.google.protobuf.AbstractParser<RegionMutation>() {
+      public RegionMutation parsePartialFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws com.google.protobuf.InvalidProtocolBufferException {
-        return new MultiAction(input, extensionRegistry);
+        return new RegionMutation(input, extensionRegistry);
       }
     };
 
     @java.lang.Override
-    public com.google.protobuf.Parser<MultiAction> getParserForType() {
+    public com.google.protobuf.Parser<RegionMutation> getParserForType() {
       return PARSER;
     }
 
     private int bitField0_;
-    // optional .MutationProto mutation = 1;
-    public static final int MUTATION_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto mutation_;
+    // required .RegionSpecifier region = 1;
+    public static final int REGION_FIELD_NUMBER = 1;
+    private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_;
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>required .RegionSpecifier region = 1;</code>
      */
-    public boolean hasMutation() {
+    public boolean hasRegion() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>required .RegionSpecifier region = 1;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation() {
-      return mutation_;
+    public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
+      return region_;
     }
     /**
-     * <code>optional .MutationProto mutation = 1;</code>
+     * <code>required .RegionSpecifier region = 1;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder() {
-      return mutation_;
+    public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
+      return region_;
     }
 
-    // optional .Get get = 2;
-    public static final int GET_FIELD_NUMBER = 2;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get get_;
+    // optional bool atomic = 2;
+    public static final int ATOMIC_FIELD_NUMBER = 2;
+    private boolean atomic_;
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>optional bool atomic = 2;</code>
+     *
+     * <pre>
+     * When set, run mutations as atomic unit.
+     * </pre>
      */
-    public boolean hasGet() {
+    public boolean hasAtomic() {
       return ((bitField0_ & 0x00000002) == 0x00000002);
     }
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>optional bool atomic = 2;</code>
+     *
+     * <pre>
+     * When set, run mutations as atomic unit.
+     * </pre>
+     */
+    public boolean getAtomic() {
+      return atomic_;
+    }
+
+    // repeated .MutationProto mutation = 3;
+    public static final int MUTATION_FIELD_NUMBER = 3;
+    private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutation_;
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
+     */
+    public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationList() {
+      return mutation_;
+    }
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
+     */
+    public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> 
+        getMutationOrBuilderList() {
+      return mutation_;
+    }
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
+     */
+    public int getMutationCount() {
+      return mutation_.size();
+    }
+    /**
+     * <code>repeated .MutationProto mutation = 3;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get getGet() {
-      return get_;
+    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation(int index) {
+      return mutation_.get(index);
     }
     /**
-     * <code>optional .Get get = 2;</code>
+     * <code>repeated .MutationProto mutation = 3;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder getGetOrBuilder() {
-      return get_;
+    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder(
+        int index) {
+      return mutation_.get(index);
     }
 
     private void initFields() {
-      mutation_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance();
-      get_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.getDefaultInstance();
+      region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
+      atomic_ = false;
+      mutation_ = java.util.Collections.emptyList();
     }
     private byte memoizedIsInitialized = -1;
     public final boolean isInitialized() {
       byte isInitialized = memoizedIsInitialized;
       if (isInitialized != -1) return isInitialized == 1;
 
-      if (hasMutation()) {
-        if (!getMutation().isInitialized()) {
-          memoizedIsInitialized = 0;
-          return false;
-        }
+      if (!hasRegion()) {
+        memoizedIsInitialized = 0;
+        return false;
+      }
+      if (!getRegion().isInitialized()) {
+        memoizedIsInitialized = 0;
+        return false;
       }
-      if (hasGet()) {
-        if (!getGet().isInitialized()) {
+      for (int i = 0; i < getMutationCount(); i++) {
+        if (!getMutation(i).isInitialized()) {
           memoizedIsInitialized = 0;
           return false;
         }
@@ -23894,10 +23966,13 @@ public final class ClientProtos {
                         throws java.io.IOException {
       getSerializedSize();
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, mutation_);
+        output.writeMessage(1, region_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
-        output.writeMessage(2, get_);
+        output.writeBool(2, atomic_);
+      }
+      for (int i = 0; i < mutation_.size(); i++) {
+        output.writeMessage(3, mutation_.get(i));
       }
       getUnknownFields().writeTo(output);
     }
@@ -23910,11 +23985,15 @@ public final class ClientProtos {
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, mutation_);
+          .computeMessageSize(1, region_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(2, get_);
+          .computeBoolSize(2, atomic_);
+      }
+      for (int i = 0; i < mutation_.size(); i++) {
+        size += com.google.protobuf.CodedOutputStream
+          .computeMessageSize(3, mutation_.get(i));
       }
       size += getUnknownFields().getSerializedSize();
       memoizedSerializedSize = size;
@@ -23933,22 +24012,24 @@ public final class ClientProtos {
       if (obj == this) {
        return true;
       }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction)) {
+      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation)) {
         return super.equals(obj);
       }
-      org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction) obj;
+      org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation) obj;
 
       boolean result = true;
-      result = result && (hasMutation() == other.hasMutation());
-      if (hasMutation()) {
-        result = result && getMutation()
-            .equals(other.getMutation());
-      }
-      result = result && (hasGet() == other.hasGet());
-      if (hasGet()) {
-        result = result && getGet()
-            .equals(other.getGet());
+      result = result && (hasRegion() == other.hasRegion());
+      if (hasRegion()) {
+        result = result && getRegion()
+            .equals(other.getRegion());
+      }
+      result = result && (hasAtomic() == other.hasAtomic());
+      if (hasAtomic()) {
+        result = result && (getAtomic()
+            == other.getAtomic());
       }
+      result = result && getMutationList()
+          .equals(other.getMutationList());
       result = result &&
           getUnknownFields().equals(other.getUnknownFields());
       return result;
@@ -23962,66 +24043,70 @@ public final class ClientProtos {
       }
       int hash = 41;
       hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasMutation()) {
-        hash = (37 * hash) + MUTATION_FIELD_NUMBER;
-        hash = (53 * hash) + getMutation().hashCode();
+      if (hasRegion()) {
+        hash = (37 * hash) + REGION_FIELD_NUMBER;
+        hash = (53 * hash) + getRegion().hashCode();
+      }
+      if (hasAtomic()) {
+        hash = (37 * hash) + ATOMIC_FIELD_NUMBER;
+        hash = (53 * hash) + hashBoolean(getAtomic());
       }
-      if (hasGet()) {
-        hash = (37 * hash) + GET_FIELD_NUMBER;
-        hash = (53 * hash) + getGet().hashCode();
+      if (getMutationCount() > 0) {
+        hash = (37 * hash) + MUTATION_FIELD_NUMBER;
+        hash = (53 * hash) + getMutationList().hashCode();
       }
       hash = (29 * hash) + getUnknownFields().hashCode();
       memoizedHashCode = hash;
       return hash;
     }
 
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         com.google.protobuf.ByteString data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(byte[] data)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         byte[] data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(java.io.InputStream input)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(java.io.InputStream input)
         throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
       return PARSER.parseFrom(input, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseDelimitedFrom(java.io.InputStream input)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseDelimitedFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
@@ -24030,7 +24115,7 @@ public final class ClientProtos {
 
     public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction prototype) {
+    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation prototype) {
       return newBuilder().mergeFrom(prototype);
     }
     public Builder toBuilder() { return newBuilder(this); }
@@ -24042,30 +24127,29 @@ public final class ClientProtos {
       return builder;
     }
     /**
-     * Protobuf type {@code MultiAction}
+     * Protobuf type {@code RegionMutation}
      *
      * <pre>
      **
-     * An action that is part of MultiRequest.
-     * This is a union type - exactly one of the fields will be set.
+     * Mutations to run against a Region.
      * </pre>
      */
     public static final class Builder extends
         com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiActionOrBuilder {
+       implements org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutationOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_MultiAction_descriptor;
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_RegionMutation_descriptor;
       }
 
       protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
           internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_MultiAction_fieldAccessorTable
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_RegionMutation_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.Builder.class);
+                org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.Builder.class);
       }
 
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.newBuilder()
+      // Construct using org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.newBuilder()
       private Builder() {
         maybeForceBuilderInitialization();
       }
@@ -24077,8 +24161,8 @@ public final class ClientProtos {
       }
       private void maybeForceBuilderInitialization() {
         if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
+          getRegionFieldBuilder();
           getMutationFieldBuilder();
-          getGetFieldBuilder();
         }
       }
       private static Builder create() {
@@ -24087,18 +24171,20 @@ public final class ClientProtos {
 
       public Builder clear() {
         super.clear();
-        if (mutationBuilder_ == null) {
-          mutation_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance();
+        if (regionBuilder_ == null) {
+          region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
         } else {
-          mutationBuilder_.clear();
+          regionBuilder_.clear();
         }
         bitField0_ = (bitField0_ & ~0x00000001);
-        if (getBuilder_ == null) {
-          get_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.getDefaultInstance();
+        atomic_ = false;
+        bitField0_ = (bitField0_ & ~0x00000002);
+        if (mutationBuilder_ == null) {
+          mutation_ = java.util.Collections.emptyList();
+          bitField0_ = (bitField0_ & ~0x00000004);
         } else {
-          getBuilder_.clear();
+          mutationBuilder_.clear();
         }
-        bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
 
@@ -24108,40 +24194,45 @@ public final class ClientProtos {
 
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_MultiAction_descriptor;
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_RegionMutation_descriptor;
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.getDefaultInstance();
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation getDefaultInstanceForType() {
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.getDefaultInstance();
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction build() {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction result = buildPartial();
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation build() {
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation result = buildPartial();
         if (!result.isInitialized()) {
           throw newUninitializedMessageException(result);
         }
         return result;
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction result = new org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction(this);
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation buildPartial() {
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation result = new org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation(this);
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
         if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
           to_bitField0_ |= 0x00000001;
         }
-        if (mutationBuilder_ == null) {
-          result.mutation_ = mutation_;
+        if (regionBuilder_ == null) {
+          result.region_ = region_;
         } else {
-          result.mutation_ = mutationBuilder_.build();
+          result.region_ = regionBuilder_.build();
         }
         if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
           to_bitField0_ |= 0x00000002;
         }
-        if (getBuilder_ == null) {
-          result.get_ = get_;
+        result.atomic_ = atomic_;
+        if (mutationBuilder_ == null) {
+          if (((bitField0_ & 0x00000004) == 0x00000004)) {
+            mutation_ = java.util.Collections.unmodifiableList(mutation_);
+            bitField0_ = (bitField0_ & ~0x00000004);
+          }
+          result.mutation_ = mutation_;
         } else {
-          result.get_ = getBuilder_.build();
+          result.mutation_ = mutationBuilder_.build();
         }
         result.bitField0_ = to_bitField0_;
         onBuilt();
@@ -24149,35 +24240,63 @@ public final class ClientProtos {
       }
 
       public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction)other);
+        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation) {
+          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation)other);
         } else {
           super.mergeFrom(other);
           return this;
         }
       }
 
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction.getDefaultInstance()) return this;
-        if (other.hasMutation()) {
-          mergeMutation(other.getMutation());
+      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation other) {
+        if (other == org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation.getDefaultInstance()) return this;
+        if (other.hasRegion()) {
+          mergeRegion(other.getRegion());
+        }
+        if (other.hasAtomic()) {
+          setAtomic(other.getAtomic());
         }
-        if (other.hasGet()) {
-          mergeGet(other.getGet());
+        if (mutationBuilder_ == null) {
+          if (!other.mutation_.isEmpty()) {
+            if (mutation_.isEmpty()) {
+              mutation_ = other.mutation_;
+              bitField0_ = (bitField0_ & ~0x00000004);
+            } else {
+              ensureMutationIsMutable();
+              mutation_.addAll(other.mutation_);
+            }
+            onChanged();
+          }
+        } else {
+          if (!other.mutation_.isEmpty()) {
+            if (mutationBuilder_.isEmpty()) {
+              mutationBuilder_.dispose();
+              mutationBuilder_ = null;
+              mutation_ = other.mutation_;
+              bitField0_ = (bitField0_ & ~0x00000004);
+              mutationBuilder_ = 
+                com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders ?
+                   getMutationFieldBuilder() : null;
+            } else {
+              mutationBuilder_.addAllMessages(other.mutation_);
+            }
+          }
         }
         this.mergeUnknownFields(other.getUnknownFields());
         return this;
       }
 
       public final boolean isInitialized() {
-        if (hasMutation()) {
-          if (!getMutation().isInitialized()) {
-            
-            return false;
-          }
+        if (!hasRegion()) {
+          
+          return false;
+        }
+        if (!getRegion().isInitialized()) {
+          
+          return false;
         }
-        if (hasGet()) {
-          if (!getGet().isInitialized()) {
+        for (int i = 0; i < getMutationCount(); i++) {
+          if (!getMutation(i).isInitialized()) {
             
             return false;
           }
@@ -24189,11 +24308,11 @@ public final class ClientProtos {
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction parsedMessage = null;
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation parsedMessage = null;
         try {
           parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MultiAction) e.getUnfinishedMessage();
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.RegionMutation) e.getUnfinishedMessage();
           throw e;
         } finally {
           if (parsedMessage != null) {
@@ -24204,267 +24323,439 @@ public final class ClientProtos {
       }
       private int bitField0_;
 
-      // optional .MutationProto mutation = 1;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto mutation_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance();
+      // required .RegionSpecifier region = 1;
+      private org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
       private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> mutationBuilder_;
+          org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> regionBuilder_;
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public boolean hasMutation() {
+      public boolean hasRegion() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation() {
-        if (mutationBuilder_ == null) {
-          return mutation_;
+      public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier getRegion() {
+        if (regionBuilder_ == null) {
+          return region_;
         } else {
-          return mutationBuilder_.getMessage();
+          return regionBuilder_.getMessage();
         }
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public Builder setMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto value) {
-        if (mutationBuilder_ == null) {
+      public Builder setRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+        if (regionBuilder_ == null) {
           if (value == null) {
             throw new NullPointerException();
           }
-          mutation_ = value;
+          region_ = value;
           onChanged();
         } else {
-          mutationBuilder_.setMessage(value);
+          regionBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public Builder setMutation(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builderForValue) {
-        if (mutationBuilder_ == null) {
-          mutation_ = builderForValue.build();
+      public Builder setRegion(
+          org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder builderForValue) {
+        if (regionBuilder_ == null) {
+          region_ = builderForValue.build();
           onChanged();
         } else {
-          mutationBuilder_.setMessage(builderForValue.build());
+          regionBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public Builder mergeMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto value) {
-        if (mutationBuilder_ == null) {
+      public Builder mergeRegion(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier value) {
+        if (regionBuilder_ == null) {
           if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              mutation_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance()) {
-            mutation_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.newBuilder(mutation_).mergeFrom(value).buildPartial();
+              region_ != org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance()) {
+            region_ =
+              org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.newBuilder(region_).mergeFrom(value).buildPartial();
           } else {
-            mutation_ = value;
+            region_ = value;
           }
           onChanged();
         } else {
-          mutationBuilder_.mergeFrom(value);
+          regionBuilder_.mergeFrom(value);
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public Builder clearMutation() {
-        if (mutationBuilder_ == null) {
-          mutation_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance();
+      public Builder clearRegion() {
+        if (regionBuilder_ == null) {
+          region_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.getDefaultInstance();
           onChanged();
         } else {
-          mutationBuilder_.clear();
+          regionBuilder_.clear();
         }
         bitField0_ = (bitField0_ & ~0x00000001);
         return this;
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder getMutationBuilder() {
+      public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder getRegionBuilder() {
         bitField0_ |= 0x00000001;
         onChanged();
-        return getMutationFieldBuilder().getBuilder();
+        return getRegionFieldBuilder().getBuilder();
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder() {
-        if (mutationBuilder_ != null) {
-          return mutationBuilder_.getMessageOrBuilder();
+      public org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder getRegionOrBuilder() {
+        if (regionBuilder_ != null) {
+          return regionBuilder_.getMessageOrBuilder();
         } else {
-          return mutation_;
+          return region_;
         }
       }
       /**
-       * <code>optional .MutationProto mutation = 1;</code>
+       * <code>required .RegionSpecifier region = 1;</code>
        */
       private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> 
-          getMutationFieldBuilder() {
-        if (mutationBuilder_ == null) {
-          mutationBuilder_ = new com.google.protobuf.SingleFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder>(
-                  mutation_,
+          org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder> 
+          getRegionFieldBuilder() {
+        if (regionBuilder_ == null) {
+          regionBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+              org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifier.Builder, org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.RegionSpecifierOrBuilder>(
+                  region_,
                   getParentForChildren(),
                   isClean());
-          mutation_ = null;
+          region_ = null;
         }
-        return mutationBuilder_;
+        return regionBuilder_;
       }
 
-      // optional .Get get = 2;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get get_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.getDefaultInstance();
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder> getBuilder_;
+      // optional bool atomic = 2;
+      private boolean atomic_ ;
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>optional bool atomic = 2;</code>
+       *
+       * <pre>
+       * When set, run mutations as atomic unit.
+       * </pre>
        */
-      public boolean hasGet() {
+      public boolean hasAtomic() {
         return ((bitField0_ & 0x00000002) == 0x00000002);
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>optional bool atomic = 2;</code>
+       *
+       * <pre>
+       * When set, run mutations as atomic unit.
+       * </pre>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get getGet() {
-        if (getBuilder_ == null) {
-          return get_;
-        } else {
-          return getBuilder_.getMessage();
-        }
+      public boolean getAtomic() {
+        return atomic_;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>optional bool atomic = 2;</code>
+       *
+       * <pre>
+       * When set, run mutations as atomic unit.
+       * </pre>
        */
-      public Builder setGet(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get value) {
-        if (getBuilder_ == null) {
-          if (value == null) {
-            throw new NullPointerException();
-          }
-          get_ = value;
-          onChanged();
-        } else {
-          getBuilder_.setMessage(value);
-        }
+      public Builder setAtomic(boolean value) {
         bitField0_ |= 0x00000002;
+        atomic_ = value;
+        onChanged();
         return this;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>optional bool atomic = 2;</code>
+       *
+       * <pre>
+       * When set, run mutations as atomic unit.
+       * </pre>
        */
-      public Builder setGet(
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder builderForValue) {
-        if (getBuilder_ == null) {
-          get_ = builderForValue.build();
-          onChanged();
-        } else {
-          getBuilder_.setMessage(builderForValue.build());
-        }
-        bitField0_ |= 0x00000002;
+      public Builder clearAtomic() {
+        bitField0_ = (bitField0_ & ~0x00000002);
+        atomic_ = false;
+        onChanged();
         return this;
       }
-      /**
-       * <code>optional .Get get = 2;</code>
-       */
-      public Builder mergeGet(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get value) {
-        if (getBuilder_ == null) {
-          if (((bitField0_ & 0x00000002) == 0x00000002) &&
-              get_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.getDefaultInstance()) {
-            get_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.newBuilder(get_).mergeFrom(value).buildPartial();
-          } else {
-            get_ = value;
+
+      // repeated .MutationProto mutation = 3;
+      private java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> mutation_ =
+        java.util.Collections.emptyList();
+      private void ensureMutationIsMutable() {
+        if (!((bitField0_ & 0x00000004) == 0x00000004)) {
+          mutation_ = new java.util.ArrayList<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto>(mutation_);
+          bitField0_ |= 0x00000004;
+         }
+      }
+
+      private com.google.protobuf.RepeatedFieldBuilder<
+          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> mutationBuilder_;
+
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> getMutationList() {
+        if (mutationBuilder_ == null) {
+          return java.util.Collections.unmodifiableList(mutation_);
+        } else {
+          return mutationBuilder_.getMessageList();
+        }
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public int getMutationCount() {
+        if (mutationBuilder_ == null) {
+          return mutation_.size();
+        } else {
+          return mutationBuilder_.getCount();
+        }
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto getMutation(int index) {
+        if (mutationBuilder_ == null) {
+          return mutation_.get(index);
+        } else {
+          return mutationBuilder_.getMessage(index);
+        }
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public Builder setMutation(
+          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto value) {
+        if (mutationBuilder_ == null) {
+          if (value == null) {
+            throw new NullPointerException();
           }
+          ensureMutationIsMutable();
+          mutation_.set(index, value);
           onChanged();
         } else {
-          getBuilder_.mergeFrom(value);
+          mutationBuilder_.setMessage(index, value);
         }
-        bitField0_ |= 0x00000002;
         return this;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>repeated .MutationProto mutation = 3;</code>
        */
-      public Builder clearGet() {
-        if (getBuilder_ == null) {
-          get_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.getDefaultInstance();
+      public Builder setMutation(
+          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builderForValue) {
+        if (mutationBuilder_ == null) {
+          ensureMutationIsMutable();
+          mutation_.set(index, builderForValue.build());
           onChanged();
         } else {
-          getBuilder_.clear();
+          mutationBuilder_.setMessage(index, builderForValue.build());
         }
-        bitField0_ = (bitField0_ & ~0x00000002);
         return this;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>repeated .MutationProto mutation = 3;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder getGetBuilder() {
-        bitField0_ |= 0x00000002;
-        onChanged();
-        return getGetFieldBuilder().getBuilder();
+      public Builder addMutation(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto value) {
+        if (mutationBuilder_ == null) {
+          if (value == null) {
+            throw new NullPointerException();
+          }
+          ensureMutationIsMutable();
+          mutation_.add(value);
+          onChanged();
+        } else {
+          mutationBuilder_.addMessage(value);
+        }
+        return this;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>repeated .MutationProto mutation = 3;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder getGetOrBuilder() {
-        if (getBuilder_ != null) {
-          return getBuilder_.getMessageOrBuilder();
+      public Builder addMutation(
+          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto value) {
+        if (mutationBuilder_ == null) {
+          if (value == null) {
+            throw new NullPointerException();
+          }
+          ensureMutationIsMutable();
+          mutation_.add(index, value);
+          onChanged();
         } else {
-          return get_;
+          mutationBuilder_.addMessage(index, value);
         }
+        return this;
       }
       /**
-       * <code>optional .Get get = 2;</code>
+       * <code>repeated .MutationProto mutation = 3;</code>
        */
-      private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder> 
-          getGetFieldBuilder() {
-        if (getBuilder_ == null) {
-          getBuilder_ = new com.google.protobuf.SingleFieldBuilder<
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Get.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.GetOrBuilder>(
-                  get_,
+      public Builder addMutation(
+          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builderForValue) {
+        if (mutationBuilder_ == null) {
+          ensureMutationIsMutable();
+          mutation_.add(builderForValue.build());
+          onChanged();
+        } else {
+          mutationBuilder_.addMessage(builderForValue.build());
+        }
+        return this;
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public Builder addMutation(
+          int index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder builderForValue) {
+        if (mutationBuilder_ == null) {
+          ensureMutationIsMutable();
+          mutation_.add(index, builderForValue.build());
+          onChanged();
+        } else {
+          mutationBuilder_.addMessage(index, builderForValue.build());
+        }
+        return this;
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public Builder addAllMutation(
+          java.lang.Iterable<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto> values) {
+        if (mutationBuilder_ == null) {
+          ensureMutationIsMutable();
+          super.addAll(values, mutation_);
+          onChanged();
+        } else {
+          mutationBuilder_.addAllMessages(values);
+        }
+        return this;
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public Builder clearMutation() {
+        if (mutationBuilder_ == null) {
+          mutation_ = java.util.Collections.emptyList();
+          bitField0_ = (bitField0_ & ~0x00000004);
+          onChanged();
+        } else {
+          mutationBuilder_.clear();
+        }
+        return this;
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public Builder removeMutation(int index) {
+        if (mutationBuilder_ == null) {
+          ensureMutationIsMutable();
+          mutation_.remove(index);
+          onChanged();
+        } else {
+          mutationBuilder_.remove(index);
+        }
+        return this;
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder getMutationBuilder(
+          int index) {
+        return getMutationFieldBuilder().getBuilder(index);
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder getMutationOrBuilder(
+          int index) {
+        if (mutationBuilder_ == null) {
+          return mutation_.get(index);  } else {
+          return mutationBuilder_.getMessageOrBuilder(index);
+        }
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public java.util.List<? extends org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> 
+           getMutationOrBuilderList() {
+        if (mutationBuilder_ != null) {
+          return mutationBuilder_.getMessageOrBuilderList();
+        } else {
+          return java.util.Collections.unmodifiableList(mutation_);
+        }
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder addMutationBuilder() {
+        return getMutationFieldBuilder().addBuilder(
+            org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance());
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder addMutationBuilder(
+          int index) {
+        return getMutationFieldBuilder().addBuilder(
+            index, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.getDefaultInstance());
+      }
+      /**
+       * <code>repeated .MutationProto mutation = 3;</code>
+       */
+      public java.util.List<org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder> 
+           getMutationBuilderList() {
+        return getMutationFieldBuilder().getBuilderList();
+      }
+      private com.google.protobuf.RepeatedFieldBuilder<
+          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder> 
+          getMutationFieldBuilder() {
+        if (mutationBuilder_ == null) {
+          mutationBuilder_ = new com.google.protobuf.RepeatedFieldBuilder<
+              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProto.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.MutationProtoOrBuilder>(
+                  mutation_,
+                  ((bitField0_ & 0x00000004) == 0x00000004),
                   getParentForChildren(),
                   isClean());
-          get_ = null;
+          mutation_ = null;
         }
-        return getBuilder_;
+        return mutationBuilder_;
       }
 
-      // @@protoc_insertion_point(builder_scope:MultiAction)
+      // @@protoc_insertion_point(builder_scope:RegionMutation)
     }
 
     static {
-      defaultInstance = new MultiAction(true);
+      defaultInstance = new RegionMutation(true);
       defaultInstance.initFields();
     }
 
-    // @@protoc_insertion_point(class_scope:MultiAction)
+    // @@protoc_insertion_point(class_scope:RegionMutation)
   }
 
-  public interface ActionResultOrBuilder
+  public interface ResultOrExceptionOrBuilder
       extends com.google.protobuf.MessageOrBuilder {
 
-    // optional .Result value = 1;
+    // optional .Result result = 1;
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    boolean hasValue();
+    boolean hasResult();
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getValue();
+    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getResult();
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getValueOrBuilder();
+    org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getResultOrBuilder();
 
     // optional .NameBytesPair exception = 2;
     /**
@@ -24481,34 +24772,31 @@ public final class ClientProtos {
     org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPairOrBuilder getExceptionOrBuilder();
   }
   /**
-   * Protobuf type {@code ActionResult}
+   * Protobuf type {@code ResultOrException}
    *
    * <pre>
    **
-   * An individual action result. The result will in the
-   * same order as the action in the request. If an action
-   * returns a value, it is set in value field. If it doesn't
-   * return anything, the result will be empty. If an action
-   * fails to execute due to any exception, the exception
-   * is returned as a stringified parameter.
+   * Either a Result or an Exception NameBytesPair (keyed by
+   * exception name whose value is the exception stringified)
+   * or maybe empty if no result and no exception.
    * </pre>
    */
-  public static final class ActionResult extends
+  public static final class ResultOrException extends
       com.google.protobuf.GeneratedMessage
-      implements ActionResultOrBuilder {
-    // Use ActionResult.newBuilder() to construct.
-    private ActionResult(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
+      implements ResultOrExceptionOrBuilder {
+    // Use ResultOrException.newBuilder() to construct.
+    private ResultOrException(com.google.protobuf.GeneratedMessage.Builder<?> builder) {
       super(builder);
       this.unknownFields = builder.getUnknownFields();
     }
-    private ActionResult(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
+    private ResultOrException(boolean noInit) { this.unknownFields = com.google.protobuf.UnknownFieldSet.getDefaultInstance(); }
 
-    private static final ActionResult defaultInstance;
-    public static ActionResult getDefaultInstance() {
+    private static final ResultOrException defaultInstance;
+    public static ResultOrException getDefaultInstance() {
       return defaultInstance;
     }
 
-    public ActionResult getDefaultInstanceForType() {
+    public ResultOrException getDefaultInstanceForType() {
       return defaultInstance;
     }
 
@@ -24518,7 +24806,7 @@ public final class ClientProtos {
         getUnknownFields() {
       return this.unknownFields;
     }
-    private ActionResult(
+    private ResultOrException(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
@@ -24544,12 +24832,12 @@ public final class ClientProtos {
             case 10: {
               org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder subBuilder = null;
               if (((bitField0_ & 0x00000001) == 0x00000001)) {
-                subBuilder = value_.toBuilder();
+                subBuilder = result_.toBuilder();
               }
-              value_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.PARSER, extensionRegistry);
+              result_ = input.readMessage(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.PARSER, extensionRegistry);
               if (subBuilder != null) {
-                subBuilder.mergeFrom(value_);
-                value_ = subBuilder.buildPartial();
+                subBuilder.mergeFrom(result_);
+                result_ = subBuilder.buildPartial();
               }
               bitField0_ |= 0x00000001;
               break;
@@ -24581,52 +24869,52 @@ public final class ClientProtos {
     }
     public static final com.google.protobuf.Descriptors.Descriptor
         getDescriptor() {
-      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ActionResult_descriptor;
+      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultOrException_descriptor;
     }
 
     protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
         internalGetFieldAccessorTable() {
-      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ActionResult_fieldAccessorTable
+      return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultOrException_fieldAccessorTable
           .ensureFieldAccessorsInitialized(
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.Builder.class);
+              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.Builder.class);
     }
 
-    public static com.google.protobuf.Parser<ActionResult> PARSER =
-        new com.google.protobuf.AbstractParser<ActionResult>() {
-      public ActionResult parsePartialFrom(
+    public static com.google.protobuf.Parser<ResultOrException> PARSER =
+        new com.google.protobuf.AbstractParser<ResultOrException>() {
+      public ResultOrException parsePartialFrom(
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws com.google.protobuf.InvalidProtocolBufferException {
-        return new ActionResult(input, extensionRegistry);
+        return new ResultOrException(input, extensionRegistry);
       }
     };
 
     @java.lang.Override
-    public com.google.protobuf.Parser<ActionResult> getParserForType() {
+    public com.google.protobuf.Parser<ResultOrException> getParserForType() {
       return PARSER;
     }
 
     private int bitField0_;
-    // optional .Result value = 1;
-    public static final int VALUE_FIELD_NUMBER = 1;
-    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value_;
+    // optional .Result result = 1;
+    public static final int RESULT_FIELD_NUMBER = 1;
+    private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result result_;
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    public boolean hasValue() {
+    public boolean hasResult() {
       return ((bitField0_ & 0x00000001) == 0x00000001);
     }
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getValue() {
-      return value_;
+    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getResult() {
+      return result_;
     }
     /**
-     * <code>optional .Result value = 1;</code>
+     * <code>optional .Result result = 1;</code>
      */
-    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getValueOrBuilder() {
-      return value_;
+    public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getResultOrBuilder() {
+      return result_;
     }
 
     // optional .NameBytesPair exception = 2;
@@ -24652,7 +24940,7 @@ public final class ClientProtos {
     }
 
     private void initFields() {
-      value_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
+      result_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
       exception_ = org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.NameBytesPair.getDefaultInstance();
     }
     private byte memoizedIsInitialized = -1;
@@ -24674,7 +24962,7 @@ public final class ClientProtos {
                         throws java.io.IOException {
       getSerializedSize();
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
-        output.writeMessage(1, value_);
+        output.writeMessage(1, result_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
         output.writeMessage(2, exception_);
@@ -24690,7 +24978,7 @@ public final class ClientProtos {
       size = 0;
       if (((bitField0_ & 0x00000001) == 0x00000001)) {
         size += com.google.protobuf.CodedOutputStream
-          .computeMessageSize(1, value_);
+          .computeMessageSize(1, result_);
       }
       if (((bitField0_ & 0x00000002) == 0x00000002)) {
         size += com.google.protobuf.CodedOutputStream
@@ -24713,16 +25001,16 @@ public final class ClientProtos {
       if (obj == this) {
        return true;
       }
-      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult)) {
+      if (!(obj instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException)) {
         return super.equals(obj);
       }
-      org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult) obj;
+      org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException other = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException) obj;
 
       boolean result = true;
-      result = result && (hasValue() == other.hasValue());
-      if (hasValue()) {
-        result = result && getValue()
-            .equals(other.getValue());
+      result = result && (hasResult() == other.hasResult());
+      if (hasResult()) {
+        result = result && getResult()
+            .equals(other.getResult());
       }
       result = result && (hasException() == other.hasException());
       if (hasException()) {
@@ -24742,9 +25030,9 @@ public final class ClientProtos {
       }
       int hash = 41;
       hash = (19 * hash) + getDescriptorForType().hashCode();
-      if (hasValue()) {
-        hash = (37 * hash) + VALUE_FIELD_NUMBER;
-        hash = (53 * hash) + getValue().hashCode();
+      if (hasResult()) {
+        hash = (37 * hash) + RESULT_FIELD_NUMBER;
+        hash = (53 * hash) + getResult().hashCode();
       }
       if (hasException()) {
         hash = (37 * hash) + EXCEPTION_FIELD_NUMBER;
@@ -24755,53 +25043,53 @@ public final class ClientProtos {
       return hash;
     }
 
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         com.google.protobuf.ByteString data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         com.google.protobuf.ByteString data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(byte[] data)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(byte[] data)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         byte[] data,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws com.google.protobuf.InvalidProtocolBufferException {
       return PARSER.parseFrom(data, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(java.io.InputStream input)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(java.io.InputStream input)
         throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
       return PARSER.parseFrom(input, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseDelimitedFrom(java.io.InputStream input)
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseDelimitedFrom(java.io.InputStream input)
         throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseDelimitedFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseDelimitedFrom(
         java.io.InputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
       return PARSER.parseDelimitedFrom(input, extensionRegistry);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         com.google.protobuf.CodedInputStream input)
         throws java.io.IOException {
       return PARSER.parseFrom(input);
     }
-    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parseFrom(
+    public static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parseFrom(
         com.google.protobuf.CodedInputStream input,
         com.google.protobuf.ExtensionRegistryLite extensionRegistry)
         throws java.io.IOException {
@@ -24810,7 +25098,7 @@ public final class ClientProtos {
 
     public static Builder newBuilder() { return Builder.create(); }
     public Builder newBuilderForType() { return newBuilder(); }
-    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult prototype) {
+    public static Builder newBuilder(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException prototype) {
       return newBuilder().mergeFrom(prototype);
     }
     public Builder toBuilder() { return newBuilder(this); }
@@ -24822,34 +25110,31 @@ public final class ClientProtos {
       return builder;
     }
     /**
-     * Protobuf type {@code ActionResult}
+     * Protobuf type {@code ResultOrException}
      *
      * <pre>
      **
-     * An individual action result. The result will in the
-     * same order as the action in the request. If an action
-     * returns a value, it is set in value field. If it doesn't
-     * return anything, the result will be empty. If an action
-     * fails to execute due to any exception, the exception
-     * is returned as a stringified parameter.
+     * Either a Result or an Exception NameBytesPair (keyed by
+     * exception name whose value is the exception stringified)
+     * or maybe empty if no result and no exception.
      * </pre>
      */
     public static final class Builder extends
         com.google.protobuf.GeneratedMessage.Builder<Builder>
-       implements org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResultOrBuilder {
+       implements org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrExceptionOrBuilder {
       public static final com.google.protobuf.Descriptors.Descriptor
           getDescriptor() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ActionResult_descriptor;
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultOrException_descriptor;
       }
 
       protected com.google.protobuf.GeneratedMessage.FieldAccessorTable
           internalGetFieldAccessorTable() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ActionResult_fieldAccessorTable
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultOrException_fieldAccessorTable
             .ensureFieldAccessorsInitialized(
-                org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.Builder.class);
+                org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.class, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.Builder.class);
       }
 
-      // Construct using org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.newBuilder()
+      // Construct using org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.newBuilder()
       private Builder() {
         maybeForceBuilderInitialization();
       }
@@ -24861,7 +25146,7 @@ public final class ClientProtos {
       }
       private void maybeForceBuilderInitialization() {
         if (com.google.protobuf.GeneratedMessage.alwaysUseFieldBuilders) {
-          getValueFieldBuilder();
+          getResultFieldBuilder();
           getExceptionFieldBuilder();
         }
       }
@@ -24871,10 +25156,10 @@ public final class ClientProtos {
 
       public Builder clear() {
         super.clear();
-        if (valueBuilder_ == null) {
-          value_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
+        if (resultBuilder_ == null) {
+          result_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
         } else {
-          valueBuilder_.clear();
+          resultBuilder_.clear();
         }
         bitField0_ = (bitField0_ & ~0x00000001);
         if (exceptionBuilder_ == null) {
@@ -24892,32 +25177,32 @@ public final class ClientProtos {
 
       public com.google.protobuf.Descriptors.Descriptor
           getDescriptorForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ActionResult_descriptor;
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.internal_static_ResultOrException_descriptor;
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult getDefaultInstanceForType() {
-        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.getDefaultInstance();
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException getDefaultInstanceForType() {
+        return org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.getDefaultInstance();
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult build() {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult result = buildPartial();
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException build() {
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException result = buildPartial();
         if (!result.isInitialized()) {
           throw newUninitializedMessageException(result);
         }
         return result;
       }
 
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult buildPartial() {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult result = new org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult(this);
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException buildPartial() {
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException result = new org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException(this);
         int from_bitField0_ = bitField0_;
         int to_bitField0_ = 0;
         if (((from_bitField0_ & 0x00000001) == 0x00000001)) {
           to_bitField0_ |= 0x00000001;
         }
-        if (valueBuilder_ == null) {
-          result.value_ = value_;
+        if (resultBuilder_ == null) {
+          result.result_ = result_;
         } else {
-          result.value_ = valueBuilder_.build();
+          result.result_ = resultBuilder_.build();
         }
         if (((from_bitField0_ & 0x00000002) == 0x00000002)) {
           to_bitField0_ |= 0x00000002;
@@ -24933,18 +25218,18 @@ public final class ClientProtos {
       }
 
       public Builder mergeFrom(com.google.protobuf.Message other) {
-        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult) {
-          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult)other);
+        if (other instanceof org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException) {
+          return mergeFrom((org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException)other);
         } else {
           super.mergeFrom(other);
           return this;
         }
       }
 
-      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult other) {
-        if (other == org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult.getDefaultInstance()) return this;
-        if (other.hasValue()) {
-          mergeValue(other.getValue());
+      public Builder mergeFrom(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException other) {
+        if (other == org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException.getDefaultInstance()) return this;
+        if (other.hasResult()) {
+          mergeResult(other.getResult());
         }
         if (other.hasException()) {
           mergeException(other.getException());
@@ -24967,11 +25252,11 @@ public final class ClientProtos {
           com.google.protobuf.CodedInputStream input,
           com.google.protobuf.ExtensionRegistryLite extensionRegistry)
           throws java.io.IOException {
-        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult parsedMessage = null;
+        org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException parsedMessage = null;
         try {
           parsedMessage = PARSER.parsePartialFrom(input, extensionRegistry);
         } catch (com.google.protobuf.InvalidProtocolBufferException e) {
-          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ActionResult) e.getUnfinishedMessage();
+          parsedMessage = (org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrException) e.getUnfinishedMessage();
           throw e;
         } finally {
           if (parsedMessage != null) {
@@ -24982,121 +25267,121 @@ public final class ClientProtos {
       }
       private int bitField0_;
 
-      // optional .Result value = 1;
-      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
+      // optional .Result result = 1;
+      private org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result result_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
       private com.google.protobuf.SingleFieldBuilder<
-          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder> valueBuilder_;
+          org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder> resultBuilder_;
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public boolean hasValue() {
+      public boolean hasResult() {
         return ((bitField0_ & 0x00000001) == 0x00000001);
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getValue() {
-        if (valueBuilder_ == null) {
-          return value_;
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result getResult() {
+        if (resultBuilder_ == null) {
+          return result_;
         } else {
-          return valueBuilder_.getMessage();
+          return resultBuilder_.getMessage();
         }
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public Builder setValue(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value) {
-        if (valueBuilder_ == null) {
+      public Builder setResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value) {
+        if (resultBuilder_ == null) {
           if (value == null) {
             throw new NullPointerException();
           }
-          value_ = value;
+          result_ = value;
           onChanged();
         } else {
-          valueBuilder_.setMessage(value);
+          resultBuilder_.setMessage(value);
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public Builder setValue(
+      public Builder setResult(
           org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder builderForValue) {
-        if (valueBuilder_ == null) {
-          value_ = builderForValue.build();
+        if (resultBuilder_ == null) {
+          result_ = builderForValue.build();
           onChanged();
         } else {
-          valueBuilder_.setMessage(builderForValue.build());
+          resultBuilder_.setMessage(builderForValue.build());
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public Builder mergeValue(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value) {
-        if (valueBuilder_ == null) {
+      public Builder mergeResult(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result value) {
+        if (resultBuilder_ == null) {
           if (((bitField0_ & 0x00000001) == 0x00000001) &&
-              value_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance()) {
-            value_ =
-              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.newBuilder(value_).mergeFrom(value).buildPartial();
+              result_ != org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance()) {
+            result_ =
+              org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.newBuilder(result_).mergeFrom(value).buildPartial();
           } else {
-            value_ = value;
+            result_ = value;
           }
           onChanged();
         } else {
-          valueBuilder_.mergeFrom(value);
+          resultBuilder_.mergeFrom(value);
         }
         bitField0_ |= 0x00000001;
         return this;
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public Builder clearValue() {
-        if (valueBuilder_ == null) {
-          value_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
+      public Builder clearResult() {
+        if (resultBuilder_ == null) {
+          result_ = org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.getDefaultInstance();
           onChanged();
         } else {
-          valueBuilder_.clear();
+          resultBuilder_.clear();
         }
         bitField0_ = (bitField0_ & ~0x00000001);
         return this;
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder getValueBuilder() {
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder getResultBuilder() {
         bitField0_ |= 0x00000001;
         onChanged();
-        return getValueFieldBuilder().getBuilder();
+        return getResultFieldBuilder().getBuilder();
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
-      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getValueOrBuilder() {
-        if (valueBuilder_ != null) {
-          return valueBuilder_.getMessageOrBuilder();
+      public org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder getResultOrBuilder() {
+        if (resultBuilder_ != null) {
+          return resultBuilder_.getMessageOrBuilder();
         } else {
-          return value_;
+          return result_;
         }
       }
       /**
-       * <code>optional .Result value = 1;</code>
+       * <code>optional .Result result = 1;</code>
        */
       private com.google.protobuf.SingleFieldBuilder<
           org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder> 
-          getValueFieldBuilder() {
-        if (valueBuilder_ == null) {
-          valueBuilder_ = new com.google.protobuf.SingleFieldBuilder<
+          getResultFieldBuilder() {
+        if (resultBuilder_ == null) {
+          resultBuilder_ = new com.google.protobuf.SingleFieldBuilder<
               org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.Result.Builder, org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ResultOrBuilder>(
-                  value_,
+                  result_,
                   getParentForChildren(),
                   isClean());
-          value_ = null;
+          result_ = null;
         }
-        return valueBuilder_;
+        return resultBuilder_;
       }
 
       // optional .NameBytesPair exception = 2;
@@ -25216,97 +25501,69 @@ public final class ClientProtos {
         return exceptionBuilder_;
       }
 
-      // @@protoc_insertion_point(builder_scope:ActionResult)
+      // @@protoc_insertion_point(builder_scope:ResultOrException)
     }
 
     static {

[... 2568 lines stripped ...]


Mime
View raw message