hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From j...@apache.org
Subject svn commit: r562608 [1/2] - in /lucene/hadoop/trunk/src/contrib/hbase: ./ src/java/org/apache/hadoop/hbase/ src/java/org/apache/hadoop/hbase/filter/ src/java/org/apache/hadoop/hbase/io/ src/java/org/apache/hadoop/hbase/mapred/ src/java/org/onelab/filte...
Date Fri, 03 Aug 2007 22:39:51 GMT
Author: jimk
Date: Fri Aug  3 15:39:43 2007
New Revision: 562608

URL: http://svn.apache.org/viewvc?view=rev&rev=562608
Log:
HADOOP-1466 Clean up warnings, visibility and javadoc issues in HBase.

Works in my environment. Since no changes were made to the code aside from white space adjustment, not testing with Hudson.

Modified:
    lucene/hadoop/trunk/src/contrib/hbase/CHANGES.txt
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/BloomFilterDescriptor.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HAbstractScanner.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HBaseConfiguration.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HConnectionManager.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HLogEdit.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterInterface.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterRegionInterface.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HScannerInterface.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HServerInfo.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreFile.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreKey.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HTableDescriptor.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/InvalidColumnNameException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/LockException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/MasterNotRunningException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NoServerForRegionException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NotServingRegionException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableExistsException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableNotDisabledException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/UnknownScannerException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/WrongRegionException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/PageRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RegExpRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterSet.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/StopRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/WhileMatchRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedData.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedDataArrayWritable.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/GroupingTableMap.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableMap.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableSplit.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/BloomFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/CountingBloomFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/DynamicBloomFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Filter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Key.java
    lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/RetouchedBloomFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/StaticTestEnvironment.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/TestCompare.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/TestHLog.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/TestMergeMeta.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/TestMergeTable.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/TestToString.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/filter/TestPageRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/filter/TestRegExpRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/filter/TestRowFilterSet.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/filter/TestStopRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/filter/TestWhileMatchRowFilter.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/shell/TestHBaseShell.java
    lucene/hadoop/trunk/src/contrib/hbase/src/test/org/apache/hadoop/hbase/util/TestKeying.java

Modified: lucene/hadoop/trunk/src/contrib/hbase/CHANGES.txt
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/CHANGES.txt?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/CHANGES.txt (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/CHANGES.txt Fri Aug  3 15:39:43 2007
@@ -84,3 +84,4 @@
      (except TestHClient and HBaseShell) have been converted to use the new client
      side objects (HTable/HBaseAdmin/HConnection) instead of HClient.
  53. HADOOP-1528 HClient for multiple tables - expose close table function
+ 54. HADOOP-1466 Clean up warnings, visibility and javadoc issues in HBase.

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/BloomFilterDescriptor.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/BloomFilterDescriptor.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/BloomFilterDescriptor.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/BloomFilterDescriptor.java Fri Aug  3 15:39:43 2007
@@ -79,6 +79,7 @@
   int vectorSize;
   int nbHash;
 
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     StringBuilder value = new StringBuilder();
@@ -103,11 +104,13 @@
     return value.toString();
   }
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object obj) {
     return compareTo(obj) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = Integer.valueOf(this.filterType).hashCode();
@@ -118,18 +121,14 @@
 
   // Writable
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     filterType = in.readInt();
     vectorSize = in.readInt();
     nbHash = in.readInt();
   }
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeInt(filterType);
     out.writeInt(vectorSize);
@@ -138,9 +137,7 @@
   
   // Comparable
   
-  /* (non-Javadoc)
-   * @see java.lang.Comparable#compareTo(java.lang.Object)
-   */
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     BloomFilterDescriptor other = (BloomFilterDescriptor)o;
     int result = this.filterType - other.filterType;

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HAbstractScanner.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HAbstractScanner.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HAbstractScanner.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HAbstractScanner.java Fri Aug  3 15:39:43 2007
@@ -41,7 +41,7 @@
   static Pattern isRegexPattern =
     Pattern.compile("^.*[\\\\+|^&*$\\[\\]\\}{)(]+.*$");
   
-  // The kind of match we are doing on a column:
+  /** The kind of match we are doing on a column: */
   private static enum MATCH_TYPE {
     /** Just check the column family name */
     FAMILY_ONLY,
@@ -51,11 +51,13 @@
     SIMPLE
   }
 
-  // This class provides column matching functions that are more sophisticated
-  // than a simple string compare. There are three types of matching:
-  // 1. Match on the column family name only
-  // 2. Match on the column family + column key regex
-  // 3. Simple match: compare column family + column key literally
+  /**
+   * This class provides column matching functions that are more sophisticated
+   * than a simple string compare. There are three types of matching:
+   * 1. Match on the column family name only
+   * 2. Match on the column family + column key regex
+   * 3. Simple match: compare column family + column key literally
+   */
   private static class ColumnMatcher {
     private boolean wildCardmatch;
     private MATCH_TYPE matchType;
@@ -84,8 +86,7 @@
       }
     }
     
-    // Matching method
-    
+    /** Matching method */
     boolean matches(Text c) throws IOException {
       if(this.matchType == MATCH_TYPE.SIMPLE) {
         return c.equals(this.col);
@@ -187,16 +188,12 @@
   /** Mechanism used to shut down the whole scan */
   public abstract void close();
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.hbase.HInternalScannerInterface#isWildcardScanner()
-   */
+  /** {@inheritDoc} */
   public boolean isWildcardScanner() {
     return this.wildcardMatch;
   }
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.hbase.HInternalScannerInterface#isMultipleMatchScanner()
-   */
+  /** {@inheritDoc} */
   public boolean isMultipleMatchScanner() {
     return this.multipleMatchers;
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HBaseConfiguration.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HBaseConfiguration.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HBaseConfiguration.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HBaseConfiguration.java Fri Aug  3 15:39:43 2007
@@ -21,7 +21,11 @@
 
 import org.apache.hadoop.conf.Configuration;
 
+/**
+ * Adds HBase configuration files to a Configuration
+ */
 public class HBaseConfiguration extends Configuration {
+  /** constructor */
   public HBaseConfiguration() {
     super();
     addDefaultResource("hbase-default.xml");

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HColumnDescriptor.java Fri Aug  3 15:39:43 2007
@@ -148,16 +148,12 @@
     this.versionNumber = COLUMN_DESCRIPTOR_VERSION;
   }
   
-  /**
-   * @return    - name of column family
-   */
+  /** @return name of column family */
   public Text getName() {
     return name;
   }
   
-  /**
-   * @return    - compression type being used for the column family
-   */
+  /** @return compression type being used for the column family */
   public CompressionType getCompression() {
     CompressionType value = null;
 
@@ -176,13 +172,12 @@
     return value;
   }
   
-  /**
-   * @return    - maximum number of versions
-   */
+  /** @return maximum number of versions */
   public int getMaxVersions() {
     return this.maxVersions;
   }
   
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     String compression = "none";
@@ -205,11 +200,13 @@
       + (bloomFilterSpecified ? bloomFilter.toString() : "none") + ")";
   }
   
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object obj) {
     return compareTo(obj) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = this.name.hashCode();
@@ -225,10 +222,9 @@
     return result;
   }
   
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.versionNumber = in.readByte();
     this.name.readFields(in);
@@ -244,6 +240,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeByte(this.versionNumber);
     this.name.write(out);
@@ -258,10 +255,9 @@
     }
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Comparable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     // NOTE: we don't do anything with the version number yet.
     // Version numbers will come into play when we introduce an incompatible

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HConnectionManager.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HConnectionManager.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HConnectionManager.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HConnectionManager.java Fri Aug  3 15:39:43 2007
@@ -113,7 +113,8 @@
     // Known region HServerAddress.toString() -> HRegionInterface 
     private HashMap<String, HRegionInterface> servers;
 
-    /** constructor
+    /** 
+     * constructor
      * @param conf Configuration object
      */
     @SuppressWarnings("unchecked")

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HLogEdit.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HLogEdit.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HLogEdit.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HLogEdit.java Fri Aug  3 15:39:43 2007
@@ -23,50 +23,62 @@
 
 import java.io.*;
 
-/*******************************************************************************
+/**
  * A log value.
  *
  * These aren't sortable; you need to sort by the matching HLogKey.
  * The table and row are already identified in HLogKey.
  * This just indicates the column and value.
- ******************************************************************************/
+ */
 public class HLogEdit implements Writable {
   private Text column = new Text();
   private byte [] val;
   private long timestamp;
 
+  /**
+   * Default constructor used by Writable
+   */
   public HLogEdit() {
     super();
   }
 
+  /**
+   * Construct a fully initialized HLogEdit
+   * @param column column name
+   * @param bval value
+   * @param timestamp timestamp for modification
+   */
   public HLogEdit(Text column, byte [] bval, long timestamp) {
     this.column.set(column);
     this.val = bval;
     this.timestamp = timestamp;
   }
 
+  /** @return the column */
   public Text getColumn() {
     return this.column;
   }
 
+  /** @return the value */
   public byte [] getVal() {
     return this.val;
   }
 
+  /** @return the timestamp */
   public long getTimestamp() {
     return this.timestamp;
   }
 
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     return getColumn().toString() + " " + this.getTimestamp() + " " +
       new String(getVal()).trim();
   }
   
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     this.column.write(out);
     out.writeShort(this.val.length);
@@ -74,6 +86,7 @@
     out.writeLong(timestamp);
   }
   
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.column.readFields(in);
     this.val = new byte[in.readShort()];

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterInterface.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterInterface.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterInterface.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterInterface.java Fri Aug  3 15:39:43 2007
@@ -30,35 +30,68 @@
  * tables.
  */
 public interface HMasterInterface extends VersionedProtocol {
-  public static final long versionID = 1L; // initial version
-
-  //////////////////////////////////////////////////////////////////////////////
-  // Check to see if master is available
-  //////////////////////////////////////////////////////////////////////////////
+  /** Interface version */
+  public static final long versionID = 1L;
 
+  /** @return true if master is available */
   public boolean isMasterRunning();
   
-  //////////////////////////////////////////////////////////////////////////////
   // Admin tools would use these cmds
-  //////////////////////////////////////////////////////////////////////////////
 
+  /**
+   * Creates a new table
+   * @param desc table descriptor
+   * @throws IOException
+   */
   public void createTable(HTableDescriptor desc) throws IOException;
+
+  /**
+   * Deletes a table
+   * @param tableName
+   * @throws IOException
+   */
   public void deleteTable(Text tableName) throws IOException;
   
+  /**
+   * Adds a column to the specified table
+   * @param tableName
+   * @param column column descriptor
+   * @throws IOException
+   */
   public void addColumn(Text tableName, HColumnDescriptor column) throws IOException;
+
+  /**
+   * Deletes a column from the specified table
+   * @param tableName
+   * @param columnName
+   * @throws IOException
+   */
   public void deleteColumn(Text tableName, Text columnName) throws IOException;
   
+  /**
+   * Puts the table on-line (only needed if table has been previously taken offline)
+   * @param tableName
+   * @throws IOException
+   */
   public void enableTable(Text tableName) throws IOException;
+  
+  /**
+   * Take table offline
+   * 
+   * @param tableName
+   * @throws IOException
+   */
   public void disableTable(Text tableName) throws IOException;
   
   /**
    * Shutdown an HBase cluster.
+   * @throws IOException
    */
   public void shutdown() throws IOException;
 
-  //////////////////////////////////////////////////////////////////////////////
-  // These are the method calls of last resort when trying to find an HRegion
-  //////////////////////////////////////////////////////////////////////////////
-
+  /**
+   * Get the location of the root region
+   * @return address of server that serves the root region
+   */
   public HServerAddress findRootRegion();
 }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterRegionInterface.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterRegionInterface.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterRegionInterface.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HMasterRegionInterface.java Fri Aug  3 15:39:43 2007
@@ -22,13 +22,30 @@
 import java.io.IOException;
 import org.apache.hadoop.ipc.VersionedProtocol;
 
-/*******************************************************************************
+/**
  * HRegionServers interact with the HMasterRegionInterface to report on local 
  * goings-on and to obtain data-handling instructions from the HMaster.
- *********************************************/
+ */
 public interface HMasterRegionInterface extends VersionedProtocol {
+  /** Interface version number */
   public static final long versionID = 1L;
+  
+  /**
+   * Called when a region server first starts
+   * @param info
+   * @throws IOException
+   */
   public void regionServerStartup(HServerInfo info) throws IOException;
+  
+  /**
+   * Called to renew lease, tell master what the region server is doing and to
+   * receive new instructions from the master
+   * 
+   * @param info server's address and start code
+   * @param msgs things the region server wants to tell the master
+   * @return instructions from the master to the region server
+   * @throws IOException
+   */
   public HMsg[] regionServerReport(HServerInfo info, HMsg msgs[])
   throws IOException;
 }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HScannerInterface.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HScannerInterface.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HScannerInterface.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HScannerInterface.java Fri Aug  3 15:39:43 2007
@@ -24,11 +24,24 @@
 import java.io.*;
 import java.util.*;
 
-/*******************************************************************************
- * HScannerInterface iterates through a set of rows.  It's implemented by several classes.
- ******************************************************************************/
+/**
+ * HScannerInterface iterates through a set of rows.  It's implemented by
+ * several classes.
+ */
 public interface HScannerInterface {
+  /**
+   * Get the next set of values
+   * @param key will contain the row and timestamp upon return
+   * @param results will contain an entry for each column family member and its value
+   * @return true if data was returned
+   * @throws IOException
+   */
   public boolean next(HStoreKey key, TreeMap<Text, byte[]> results)
   throws IOException;
+  
+  /**
+   * Closes a scanner and releases any resources it has allocated
+   * @throws IOException
+   */
   public void close() throws IOException;
 }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HServerInfo.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HServerInfo.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HServerInfo.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HServerInfo.java Fri Aug  3 15:39:43 2007
@@ -23,51 +23,67 @@
 
 import java.io.*;
 
-/*******************************************************************************
- * HRSInfo contains metainfo about an HRegionServer, including details about the
- * source machine and load statistics.
- ******************************************************************************/
+/**
+ * HServerInfo contains metainfo about an HRegionServer, Currently it only
+ * contains the server start code.
+ * 
+ * In the future it will contain information about the source machine and
+ * load statistics.
+ */
 public class HServerInfo implements Writable {
   private HServerAddress serverAddress;
   private long startCode;
 
+  /** default constructor - used by Writable */
   public HServerInfo() {
     this.serverAddress = new HServerAddress();
     this.startCode = 0;
   }
   
+  /**
+   * Constructs a fully initialized object
+   * @param serverAddress
+   * @param startCode
+   */
   public HServerInfo(HServerAddress serverAddress, long startCode) {
     this.serverAddress = new HServerAddress(serverAddress);
     this.startCode = startCode;
   }
   
+  /**
+   * Construct a new object using another as input (like a copy constructor)
+   * @param other
+   */
   public HServerInfo(HServerInfo other) {
     this.serverAddress = new HServerAddress(other.getServerAddress());
     this.startCode = other.getStartCode();
   }
-  
+
+  /** @return the server address */
   public HServerAddress getServerAddress() {
     return serverAddress;
   }
-  
+ 
+  /** @return the server start code */
   public long getStartCode() {
     return startCode;
   }
   
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     return "address: " + this.serverAddress + ", startcode: " + this.startCode;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.serverAddress.readFields(in);
     this.startCode = in.readLong();
   }
 
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     this.serverAddress.write(out);
     out.writeLong(this.startCode);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreFile.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreFile.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreFile.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreFile.java Fri Aug  3 15:39:43 2007
@@ -29,14 +29,21 @@
 import java.io.*;
 import java.util.*;
 
-/*******************************************************************************
+/**
  * Each HStore maintains a bunch of different data files.
  *
+ * An HStoreFile tracks 4 things: its parent dir, the region identifier, the 
+ * column family, and the file identifier.  If you know those four things, you
+ * know how to obtain the right HStoreFile.
+ *
+ * When merging or splitting HRegions, we might want to modify one of the 
+ * params for an HStoreFile (effectively moving it elsewhere).
+ * 
  * The filename is a mix of the parent dir, the region name, the column name, 
  * and the file identifier.
  * 
  * This class handles all that path-building stuff for you.
- ******************************************************************************/
+ */
 public class HStoreFile implements HConstants, WritableComparable {
   private static final Log LOG = LogFactory.getLog(HStoreFile.class.getName());
   static final byte INFO_SEQ_NUM = 0;
@@ -53,14 +60,7 @@
   long fileId;
   Configuration conf;
 
-  /**
-   * An HStoreFile tracks 4 things: its parent dir, the region identifier, the 
-   * column family, and the file identifier.  If you know those four things, you
-   * know how to obtain the right HStoreFile.
-   *
-   * When merging or splitting HRegions, we might want to modify one of the 
-   * params for an HStoreFile (effectively moving it elsewhere).
-   */
+  /** Constructor used by Writable */
   HStoreFile(Configuration conf) {
     this.conf = conf;
     this.dir = new Path(Path.CUR_DIR);
@@ -69,6 +69,14 @@
     this.fileId = 0;
   }
   
+  /**
+   * Constructor that fully initializes the object
+   * @param conf Configuration object
+   * @param dir directory path
+   * @param regionName name of the region
+   * @param colFamily name of the column family
+   * @param fileId file identifier
+   */
   HStoreFile(Configuration conf, Path dir, Text regionName, 
       Text colFamily, long fileId) {
     
@@ -79,31 +87,35 @@
     this.fileId = fileId;
   }
 
-  // Get the individual components
-  
+  /** @return the directory path */
   Path getDir() {
     return dir;
   }
-  
+
+  /** @return the region name */
   Text getRegionName() {
     return regionName;
   }
-  
+
+  /** @return the column family */
   Text getColFamily() {
     return colFamily;
   }
-  
+
+  /** @return the file identifier */
   long fileId() {
     return fileId;
   }
 
   // Build full filenames from those components
   
+  /** @return path for MapFile */
   Path getMapFilePath() {
     return new Path(HStoreFile.getMapDir(dir, regionName, colFamily), 
         HSTORE_DATFILE_PREFIX + fileId);
   }
   
+  /** @return path for info file */
   Path getInfoFilePath() {
     return new Path(HStoreFile.getInfoDir(dir, regionName, colFamily), 
         HSTORE_INFOFILE_PREFIX + fileId);
@@ -111,34 +123,41 @@
 
   // Static methods to build partial paths to internal directories.  Useful for 
   // HStore construction and log-rebuilding.
-  
+
+  /** @return the map directory path */
   static Path getMapDir(Path dir, Text regionName, Text colFamily) {
     return new Path(dir, new Path(HREGIONDIR_PREFIX + regionName, 
         new Path(colFamily.toString(), HSTORE_DATFILE_DIR)));
   }
 
+  /** @return the info directory path */
   static Path getInfoDir(Path dir, Text regionName, Text colFamily) {
     return new Path(dir, new Path(HREGIONDIR_PREFIX + regionName, 
         new Path(colFamily.toString(), HSTORE_INFO_DIR)));
   }
-  
+
+  /** @return the bloom filter directory path */
   static Path getFilterDir(Path dir, Text regionName, Text colFamily) {
     return new Path(dir, new Path(HREGIONDIR_PREFIX + regionName,
         new Path(colFamily.toString(), HSTORE_FILTER_DIR)));
   }
 
+  /** @return the HStore directory path */
   static Path getHStoreDir(Path dir, Text regionName, Text colFamily) {
     return new Path(dir, new Path(HREGIONDIR_PREFIX + regionName, 
         colFamily.toString()));
   }
 
+  /** @return the HRegion directory path */
   static Path getHRegionDir(Path dir, Text regionName) {
     return new Path(dir, new Path(HREGIONDIR_PREFIX + regionName));
   }
 
   /**
-   * Obtain a brand-new randomly-named HStoreFile.  Checks the existing
-   * filesystem if the file already exists.
+   * @return a brand-new randomly-named HStoreFile.
+   * 
+   * Checks the filesystem to determine if the file already exists. If so, it
+   * will keep generating names until it generates a name that does not exist.
    */
   static HStoreFile obtainNewHStoreFile(Configuration conf, Path dir, 
       Text regionName, Text colFamily, FileSystem fs) throws IOException {
@@ -157,10 +176,18 @@
   }
 
   /**
-   * Create a series of HStoreFiles loaded from the given directory.
+   * Creates a series of HStoreFiles loaded from the given directory.
    * 
    * There must be a matching 'mapdir' and 'loginfo' pair of files.
    * If only one exists, we'll delete it.
+   *
+   * @param conf Configuration object
+   * @param dir directory path
+   * @param regionName region name
+   * @param colFamily column family
+   * @param fs file system
+   * @return Vector of HStoreFiles
+   * @throws IOException
    */
   static Vector<HStoreFile> loadHStoreFiles(Configuration conf, Path dir, 
       Text regionName, Text colFamily, FileSystem fs) throws IOException {
@@ -173,8 +200,12 @@
       String name = datfiles[i].getName();
       
       if(name.startsWith(HSTORE_DATFILE_PREFIX)) {
-        Long fileId = Long.parseLong(name.substring(HSTORE_DATFILE_PREFIX.length()));
-        HStoreFile curfile = new HStoreFile(conf, dir, regionName, colFamily, fileId);
+        Long fileId =
+          Long.parseLong(name.substring(HSTORE_DATFILE_PREFIX.length()));
+
+        HStoreFile curfile =
+          new HStoreFile(conf, dir, regionName, colFamily, fileId);
+
         Path mapfile = curfile.getMapFilePath();
         Path infofile = curfile.getInfoFilePath();
         
@@ -193,8 +224,12 @@
       String name = infofiles[i].getName();
       
       if(name.startsWith(HSTORE_INFOFILE_PREFIX)) {
-        long fileId = Long.parseLong(name.substring(HSTORE_INFOFILE_PREFIX.length()));
-        HStoreFile curfile = new HStoreFile(conf, dir, regionName, colFamily, fileId);
+        long fileId =
+          Long.parseLong(name.substring(HSTORE_INFOFILE_PREFIX.length()));
+
+        HStoreFile curfile =
+          new HStoreFile(conf, dir, regionName, colFamily, fileId);
+
         Path mapfile = curfile.getMapFilePath();
         
         if(! fs.exists(mapfile)) {
@@ -205,31 +240,40 @@
     return results;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // File handling
-  //////////////////////////////////////////////////////////////////////////////
 
   /**
    * Break this HStoreFile file into two new parts, which live in different 
    * brand-new HRegions.
+   *
+   * @param midKey the key which will be the starting key of the second region
+   * @param dstA the file which will contain keys from the start of the source
+   * @param dstB the file which will contain keys from midKey to end of source
+   * @param fs file system
+   * @param c configuration
+   * @throws IOException
    */
   void splitStoreFile(Text midKey, HStoreFile dstA, HStoreFile dstB,
-      FileSystem fs, Configuration c)
-  throws IOException {
+      FileSystem fs, Configuration c) throws IOException {
+    
     // Copy the appropriate tuples to one MapFile or the other.
+    
     MapFile.Reader in = new MapFile.Reader(fs, getMapFilePath().toString(), c);
     try {
       MapFile.Writer outA = new MapFile.Writer(c, fs, 
         dstA.getMapFilePath().toString(), HStoreKey.class,
         ImmutableBytesWritable.class);
+      
       try {
         MapFile.Writer outB = new MapFile.Writer(c, fs, 
           dstB.getMapFilePath().toString(), HStoreKey.class,
           ImmutableBytesWritable.class);
+        
         try {
           long count = 0;
           HStoreKey readkey = new HStoreKey();
           ImmutableBytesWritable readval = new ImmutableBytesWritable();
+          
           while(in.next(readkey, readval)) {
             if(readkey.getRow().compareTo(midKey) < 0) {
               outA.append(readkey, readval);
@@ -243,12 +287,15 @@
               }
             }
           }
+          
         } finally {
           outB.close();
         }
+        
       } finally {
         outA.close();
       }
+      
     } finally {
       in.close();
     }
@@ -260,8 +307,12 @@
   }
 
   /**
-   * Write to this HStoreFile with all the contents of the given source HStoreFiles.
-   * We are merging multiple regions into a single new one.
+   * Merges the contents of the given source HStoreFiles into a single new one.
+   *
+   * @param srcFiles files to be merged
+   * @param fs file system
+   * @param conf configuration object
+   * @throws IOException
    */
   void mergeStoreFiles(Vector<HStoreFile> srcFiles, FileSystem fs, 
       Configuration conf) throws IOException {
@@ -273,9 +324,9 @@
       HStoreKey.class, ImmutableBytesWritable.class);
     
     try {
-      for(Iterator<HStoreFile> it = srcFiles.iterator(); it.hasNext(); ) {
-        HStoreFile src = it.next();
-        MapFile.Reader in = new MapFile.Reader(fs, src.getMapFilePath().toString(), conf);
+      for(HStoreFile src: srcFiles) {
+        MapFile.Reader in =
+          new MapFile.Reader(fs, src.getMapFilePath().toString(), conf);
         
         try {
           HStoreKey readkey = new HStoreKey();
@@ -283,6 +334,7 @@
           while(in.next(readkey, readval)) {
             out.append(readkey, readval);
           }
+          
         } finally {
           in.close();
         }
@@ -293,6 +345,7 @@
     }
 
     // Build a unified InfoFile from the source InfoFiles.
+    
     long unifiedSeqId = -1;
     for(Iterator<HStoreFile> it = srcFiles.iterator(); it.hasNext(); ) {
       HStoreFile hsf = it.next();
@@ -304,7 +357,13 @@
     writeInfo(fs, unifiedSeqId);
   }
 
-  /** Read in an info file, give it a unique ID. */
+  /** 
+   * Reads in an info file, and gives it a unique ID.
+   *
+   * @param fs file system
+   * @return new unique id
+   * @throws IOException
+   */
   long loadInfo(FileSystem fs) throws IOException {
     Path p = getInfoFilePath();
     DataInputStream in = new DataInputStream(fs.open(p));
@@ -319,7 +378,13 @@
     }
   }
   
-  /** Write the file-identifier to disk */
+  /**
+   * Writes the file-identifier to disk
+   * 
+   * @param fs file system
+   * @param infonum file id
+   * @throws IOException
+   */
   void writeInfo(FileSystem fs, long infonum) throws IOException {
     Path p = getInfoFilePath();
     DataOutputStream out = new DataOutputStream(fs.create(p));
@@ -333,11 +398,13 @@
     }
   }
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = this.dir.hashCode();
@@ -347,13 +414,9 @@
     return result;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeUTF(dir.toString());
     regionName.write(out);
@@ -361,9 +424,7 @@
     out.writeLong(fileId);
   }
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.dir = new Path(in.readUTF());
     this.regionName.readFields(in);
@@ -371,13 +432,9 @@
     this.fileId = in.readLong();
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Comparable
-  //////////////////////////////////////////////////////////////////////////////
 
-  /* (non-Javadoc)
-   * @see java.lang.Comparable#compareTo(java.lang.Object)
-   */
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     HStoreFile other = (HStoreFile) o;
     int result = this.dir.compareTo(other.dir);    

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreKey.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreKey.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreKey.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HStoreKey.java Fri Aug  3 15:39:43 2007
@@ -147,16 +147,14 @@
     this.timestamp = timestamp;
   }
   
-  /**
-   * @return Approximate size in bytes of this key.
-   */
+  /** @return Approximate size in bytes of this key. */
   public long getSize() {
     return this.row.getLength() + this.column.getLength() +
       8 /* There is no sizeof in java. Presume long is 8 (64bit machine)*/;
   }
   
   /**
-   * Construct a new HStoreKey from another
+   * Constructs a new HStoreKey from another
    * 
    * @param other the source key
    */
@@ -218,6 +216,7 @@
   }
   
   /**
+   * Compares the row and column of two keys
    * @param other Key to compare against. Compares row and column.
    * @return True if same row and column.
    * @see #matchesWithoutColumn(HStoreKey)
@@ -229,6 +228,8 @@
   }
   
   /**
+   * Compares the row and timestamp of two keys
+   * 
    * @param other Key to copmare against. Compares row and timestamp.
    * 
    * @return True if same row and timestamp is greater than <code>other</code>
@@ -241,6 +242,8 @@
   }
   
   /**
+   * Compares the row and column family of two keys
+   * 
    * @param that Key to compare against. Compares row and column family
    * 
    * @return true if same row and column family
@@ -255,16 +258,19 @@
         compareTo(extractFamily(that.getColumn())) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     return row.toString() + "/" + column.toString() + "/" + timestamp;
   }
   
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object obj) {
     return compareTo(obj) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = this.row.hashCode();
@@ -273,13 +279,9 @@
     return result;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Comparable
-  //////////////////////////////////////////////////////////////////////////////
 
-  /* (non-Javadoc)
-   * @see java.lang.Comparable#compareTo(java.lang.Object)
-   */
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     HStoreKey other = (HStoreKey) o;
     int result = this.row.compareTo(other.row);
@@ -296,22 +298,16 @@
     return result;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     row.write(out);
     column.write(out);
     out.writeLong(timestamp);
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     row.readFields(in);
     column.readFields(in);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HTableDescriptor.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HTableDescriptor.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HTableDescriptor.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/HTableDescriptor.java Fri Aug  3 15:39:43 2007
@@ -78,7 +78,7 @@
   }
 
   /**
-   * Add a column family.
+   * Adds a column family.
    * @param family HColumnDescriptor of familyto add.
    */
   public void addFamily(HColumnDescriptor family) {
@@ -95,7 +95,8 @@
     return families.containsKey(family);
   }
 
-  /** All the column families in this table.
+  /** 
+   * All the column families in this table.
    * 
    *  TODO: What is this used for? Seems Dangerous to let people play with our
    *  private members.
@@ -106,16 +107,19 @@
     return families;
   }
 
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     return "name: " + this.name.toString() + ", families: " + this.families;
       }
   
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object obj) {
     return compareTo(obj) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     // TODO: Cache.
@@ -128,10 +132,9 @@
     return result;
   }
   
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     name.write(out);
     out.writeInt(families.size());
@@ -141,6 +144,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.name.readFields(in);
     int numCols = in.readInt();
@@ -152,10 +156,9 @@
     }
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Comparable
-  //////////////////////////////////////////////////////////////////////////////
 
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     HTableDescriptor other = (HTableDescriptor) o;
     int result = name.compareTo(other.name);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/InvalidColumnNameException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/InvalidColumnNameException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/InvalidColumnNameException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/InvalidColumnNameException.java Fri Aug  3 15:39:43 2007
@@ -21,12 +21,20 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown when an invalid column name is encountered
+ */
 public class InvalidColumnNameException extends IOException {
   private static final long serialVersionUID = 1L << 29 - 1L;
+  /** default constructor */
   public InvalidColumnNameException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public InvalidColumnNameException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/LockException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/LockException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/LockException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/LockException.java Fri Aug  3 15:39:43 2007
@@ -21,12 +21,20 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown when a locking error is encountered
+ */
 public class LockException extends IOException {
   private static final long serialVersionUID = 1L << 13 - 1L;
+  /** default constructor */
   public LockException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public LockException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/MasterNotRunningException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/MasterNotRunningException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/MasterNotRunningException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/MasterNotRunningException.java Fri Aug  3 15:39:43 2007
@@ -21,12 +21,20 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown if the master is not running
+ */
 public class MasterNotRunningException extends IOException {
   private static final long serialVersionUID = 1L << 23 - 1L;
+  /** default constructor */
   public MasterNotRunningException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public MasterNotRunningException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NoServerForRegionException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NoServerForRegionException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NoServerForRegionException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NoServerForRegionException.java Fri Aug  3 15:39:43 2007
@@ -21,13 +21,21 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown when no region server can be found for a region
+ */
 public class NoServerForRegionException extends IOException {
   private static final long serialVersionUID = 1L << 11 - 1L;
 
+  /** default constructor */
   public NoServerForRegionException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public NoServerForRegionException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NotServingRegionException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NotServingRegionException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NotServingRegionException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/NotServingRegionException.java Fri Aug  3 15:39:43 2007
@@ -21,12 +21,22 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown by a region server if it is sent a request for a region it is not
+ * serving.
+ */
 public class NotServingRegionException extends IOException {
   private static final long serialVersionUID = 1L << 17 - 1L;
+
+  /** default constructor */
   public NotServingRegionException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public NotServingRegionException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableExistsException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableExistsException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableExistsException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableExistsException.java Fri Aug  3 15:39:43 2007
@@ -17,11 +17,21 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown when a table exists but should not
+ */
 public class TableExistsException extends IOException {
+  private static final long serialVersionUID = 1L << 7 - 1L;
+  /** default constructor */
   public TableExistsException() {
     super();
   }
 
+  /**
+   * Constructor
+   * 
+   * @param s message
+   */
   public TableExistsException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableNotDisabledException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableNotDisabledException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableNotDisabledException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/TableNotDisabledException.java Fri Aug  3 15:39:43 2007
@@ -21,12 +21,20 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown if a table should be offline but is not
+ */
 public class TableNotDisabledException extends IOException {
   private static final long serialVersionUID = 1L << 19 - 1L;
+  /** default constructor */
   public TableNotDisabledException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public TableNotDisabledException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/UnknownScannerException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/UnknownScannerException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/UnknownScannerException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/UnknownScannerException.java Fri Aug  3 15:39:43 2007
@@ -21,13 +21,21 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown if a region server is passed an unknown scanner id
+ */
 public class UnknownScannerException extends IOException {
   private static final long serialVersionUID = 993179627856392526L;
 
+  /** constructor */
   public UnknownScannerException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public UnknownScannerException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/WrongRegionException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/WrongRegionException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/WrongRegionException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/WrongRegionException.java Fri Aug  3 15:39:43 2007
@@ -21,13 +21,21 @@
 
 import java.io.IOException;
 
+/**
+ * Thrown when a request contains a key which is not part of this region
+ */
 public class WrongRegionException extends IOException {
   private static final long serialVersionUID = 993179627856392526L;
 
+  /** constructor */
   public WrongRegionException() {
     super();
   }
 
+  /**
+   * Constructor
+   * @param s message
+   */
   public WrongRegionException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/InvalidRowFilterException.java Fri Aug  3 15:39:43 2007
@@ -25,10 +25,16 @@
 public class InvalidRowFilterException extends RuntimeException {
   private static final long serialVersionUID = 2667894046345657865L;
 
+
+  /** constructor */
   public InvalidRowFilterException() {
     super();
   }
 
+  /**
+   * constructor
+   * @param s message
+   */
   public InvalidRowFilterException(String s) {
     super(s);
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/PageRowFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/PageRowFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/PageRowFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/PageRowFilter.java Fri Aug  3 15:39:43 2007
@@ -81,11 +81,9 @@
     rowsAccepted = 0;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
-  public void rowProcessed(boolean filtered, Text rowKey) {
+  /** {@inheritDoc} */
+  public void rowProcessed(boolean filtered,
+      @SuppressWarnings("unused") Text rowKey) {
     if (!filtered) {
       this.rowsAccepted++;
       if (LOG.isDebugEnabled()) {

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RegExpRowFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RegExpRowFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RegExpRowFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RegExpRowFilter.java Fri Aug  3 15:39:43 2007
@@ -80,18 +80,13 @@
     this.setColumnFilters(columnFilter);
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
+  @SuppressWarnings("unused")
   public void rowProcessed(boolean filtered, Text rowKey) {
     //doesn't care
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean processAlways() {
     return false;
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterInterface.java Fri Aug  3 15:39:43 2007
@@ -45,6 +45,7 @@
    * RowFilterSet with an OR operator.
    * 
    * @see RowFilterSet
+   * @param filtered
    * @param key
    */
   void rowProcessed(boolean filtered, Text key);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterSet.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterSet.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterSet.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/RowFilterSet.java Fri Aug  3 15:39:43 2007
@@ -38,8 +38,12 @@
  */
 public class RowFilterSet implements RowFilterInterface {
 
+  /** set operator */
   public static enum Operator {
-    MUST_PASS_ALL, MUST_PASS_ONE
+    /** !AND */
+    MUST_PASS_ALL,
+    /** !OR */
+    MUST_PASS_ONE
   }
 
   private Operator operator = Operator.MUST_PASS_ALL;
@@ -77,10 +81,7 @@
     this.operator = operator;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void validate(final Text[] columns) {
     for (RowFilterInterface filter : filters) {
       filter.validate(columns);
@@ -91,10 +92,7 @@
     }
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void reset() {
     for (RowFilterInterface filter : filters) {
       filter.reset();
@@ -105,10 +103,7 @@
     }
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void rowProcessed(boolean filtered, Text rowKey) {
     for (RowFilterInterface filter : filters) {
       filter.rowProcessed(filtered, rowKey);
@@ -119,10 +114,7 @@
     }
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean processAlways() {
     for (RowFilterInterface filter : filters) {
       if (filter.processAlways()) {
@@ -136,10 +128,7 @@
     return false;
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filterAllRemaining() {
     boolean result = operator == Operator.MUST_PASS_ONE;
     for (RowFilterInterface filter : filters) {
@@ -167,10 +156,7 @@
     return result;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filter(final Text rowKey) {
     boolean resultFound = false;
     boolean result = operator == Operator.MUST_PASS_ONE;
@@ -205,10 +191,7 @@
     return result;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filter(final Text rowKey, final Text colKey, 
     final byte[] data) {
     boolean resultFound = false;
@@ -248,10 +231,7 @@
     return result;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filterNotNull(final TreeMap<Text, byte[]> columns) {
     boolean resultFound = false;
     boolean result = operator == Operator.MUST_PASS_ONE;
@@ -286,10 +266,7 @@
     return result;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void readFields(final DataInput in) throws IOException {
     byte opByte = in.readByte();
     operator = Operator.values()[opByte];
@@ -323,10 +300,7 @@
 
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void write(final DataOutput out) throws IOException {
     out.writeByte(operator.ordinal());
     out.writeInt(filters.size());

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/StopRowFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/StopRowFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/StopRowFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/StopRowFilter.java Fri Aug  3 15:39:43 2007
@@ -80,34 +80,23 @@
     // Nothing to reset
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
+  @SuppressWarnings("unused")
   public void rowProcessed(boolean filtered, Text rowKey) {
     // Doesn't care
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean processAlways() {
     return false;
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filterAllRemaining() {
     return false;
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filter(final Text rowKey) {
     boolean result = this.stopRowKey.compareTo(rowKey) <= 0;
     if (LOG.isDebugEnabled()) {
@@ -118,6 +107,8 @@
   }
 
   /**
+   * {@inheritDoc}
+   *
    * Because StopRowFilter does not examine column information, this method 
    * defaults to calling the rowKey-only version of filter.
    */
@@ -127,7 +118,8 @@
     return filter(rowKey);
   }
 
-  /**
+  /** {@inheritDoc}
+   *
    * Because StopRowFilter does not examine column information, this method 
    * defaults to calling filterAllRemaining().
    * 
@@ -138,18 +130,12 @@
     return filterAllRemaining();
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     stopRowKey = new Text(in.readUTF());
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeUTF(stopRowKey.toString());
   }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/WhileMatchRowFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/WhileMatchRowFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/WhileMatchRowFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/filter/WhileMatchRowFilter.java Fri Aug  3 15:39:43 2007
@@ -50,6 +50,10 @@
     super();
   }
   
+  /**
+   * Constructor
+   * @param filter
+   */
   public WhileMatchRowFilter(RowFilterInterface filter) {
     this.filter = filter;
   }
@@ -63,10 +67,7 @@
     return this.filter;
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void reset() {
     if (LOG.isDebugEnabled()) {
       LOG.debug("Resetting.");
@@ -75,10 +76,7 @@
     this.filter.reset();
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean processAlways() {
     return true;
   }
@@ -94,10 +92,7 @@
     return this.filterAllRemaining || this.filter.filterAllRemaining();
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filter(final Text rowKey) {
     changeFAR(this.filter.filter(rowKey));
     boolean result = filterAllRemaining();
@@ -107,10 +102,7 @@
     return result;
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filter(final Text rowKey, final Text colKey,
     final byte[] data) {
     changeFAR(this.filter.filter(rowKey, colKey, data));
@@ -122,10 +114,7 @@
     return result;
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public boolean filterNotNull(final TreeMap<Text, byte[]> columns) {
     changeFAR(this.filter.filterNotNull(columns));
     boolean result = filterAllRemaining();
@@ -150,26 +139,17 @@
     }
   }
 
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void rowProcessed(boolean filtered, Text rowKey) {
     this.filter.rowProcessed(filtered, rowKey);
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void validate(Text[] columns) {
     this.filter.validate(columns);
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     String className = in.readUTF();
     
@@ -193,10 +173,7 @@
     }
   }
   
-  /**
-   * 
-   * {@inheritDoc}
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeUTF(this.filter.getClass().getName());
     this.filter.write(out);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/ImmutableBytesWritable.java Fri Aug  3 15:39:43 2007
@@ -88,7 +88,7 @@
   }
   
   /**
-   * Get the current size of the buffer.
+   * @return the current size of the buffer.
    */
   public int getSize() {
     if (this.bytes == null) {
@@ -99,13 +99,13 @@
   }
 
 
-  // inherit javadoc
+  /** {@inheritDoc} */
   public void readFields(final DataInput in) throws IOException {
     this.bytes = new byte[in.readInt()];
     in.readFully(this.bytes, 0, this.bytes.length);
   }
   
-  // inherit javadoc
+  /** {@inheritDoc} */
   public void write(final DataOutput out) throws IOException {
     out.writeInt(this.bytes.length);
     out.write(this.bytes, 0, this.bytes.length);
@@ -113,6 +113,8 @@
   
   // Below methods copied from BytesWritable
   
+  /** {@inheritDoc} */
+  @Override
   public int hashCode() {
     return WritableComparator.hashBytes(bytes, this.bytes.length);
   }
@@ -127,6 +129,12 @@
     return compareTo(((ImmutableBytesWritable)right_obj).get());
   }
   
+  /**
+   * Compares the bytes in this object to the specified byte array
+   * @param that
+   * @return Positive if left is bigger than right, 0 if they are equal, and
+   *         negative if left is smaller than right.
+   */
   public int compareTo(final byte [] that) {
     int diff = this.bytes.length - that.length;
     return (diff != 0)?
@@ -135,9 +143,8 @@
         0, that.length);
   }
   
-  /**
-   * Are the two byte sequences equal?
-   */
+  /** {@inheritDoc} */
+  @Override
   public boolean equals(Object right_obj) {
     if (right_obj instanceof ImmutableBytesWritable) {
       return compareTo(right_obj) == 0;
@@ -145,9 +152,8 @@
     return false;
   }
   
-  /**
-   * Generate the stream of bytes as hex pairs separated by ' '.
-   */
+  /** {@inheritDoc} */
+  @Override
   public String toString() { 
     StringBuffer sb = new StringBuffer(3*this.bytes.length);
     for (int idx = 0; idx < this.bytes.length; idx++) {
@@ -170,14 +176,14 @@
   public static class Comparator extends WritableComparator {
     private BytesWritable.Comparator comparator =
       new BytesWritable.Comparator();
-    
+
+    /** constructor */
     public Comparator() {
       super(ImmutableBytesWritable.class);
     }
     
-    /**
-     * Compare the buffers in serialized form.
-     */
+    /** {@inheritDoc} */
+    @Override
     public int compare(byte[] b1, int s1, int l1, byte[] b2, int s2, int l2) {
       return comparator.compare(b1, s1, l1, b2, s2, l2);
     }

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedData.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedData.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedData.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedData.java Fri Aug  3 15:39:43 2007
@@ -56,22 +56,16 @@
     return data;
   }
 
-  //////////////////////////////////////////////////////////////////////////////
   // Writable
-  //////////////////////////////////////////////////////////////////////////////
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     key.write(out);
     out.writeInt(this.data.length);
     out.write(this.data);
   }
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     key.readFields(in);
     this.data = new byte[in.readInt()];

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedDataArrayWritable.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedDataArrayWritable.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedDataArrayWritable.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/io/KeyedDataArrayWritable.java Fri Aug  3 15:39:43 2007
@@ -61,6 +61,7 @@
 
   // Writable
   
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     int len = in.readInt();
     m_data = new KeyedData[len];
@@ -70,6 +71,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     int len = m_data.length;
     out.writeInt(len);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/GroupingTableMap.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/GroupingTableMap.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/GroupingTableMap.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/GroupingTableMap.java Fri Aug  3 15:39:43 2007
@@ -66,9 +66,7 @@
     job.set(GROUP_COLUMNS, groupColumns);
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.hbase.mapred.TableMap#configure(org.apache.hadoop.mapred.JobConf)
-   */
+  /** {@inheritDoc} */
   @Override
   public void configure(JobConf job) {
     super.configure(job);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableMap.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableMap.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableMap.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableMap.java Fri Aug  3 15:39:43 2007
@@ -70,9 +70,7 @@
     job.set(TableInputFormat.COLUMN_LIST, columns);
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.mapred.MapReduceBase#configure(org.apache.hadoop.mapred.JobConf)
-   */
+  /** {@inheritDoc} */
   @Override
   public void configure(JobConf job) {
     super.configure(job);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableSplit.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableSplit.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableSplit.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/apache/hadoop/hbase/mapred/TableSplit.java Fri Aug  3 15:39:43 2007
@@ -69,40 +69,33 @@
     return m_endRow;
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.mapred.InputSplit#getLength()
-   */
+  /** {@inheritDoc} */
   public long getLength() {
     // Not clear how to obtain this... seems to be used only for sorting splits
     return 0;
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.mapred.InputSplit#getLocations()
-   */
+  /** {@inheritDoc} */
   public String[] getLocations() {
     // Return a random node from the cluster for now
     return new String[] { };
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     m_tableName.readFields(in);
     m_startRow.readFields(in);
     m_endRow.readFields(in);
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     m_tableName.write(out);
     m_startRow.write(out);
     m_endRow.write(out);
   }
 
+  /** {@inheritDoc} */
   @Override
   public String toString() {
     return m_tableName +"," + m_startRow + "," + m_endRow;

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/BloomFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/BloomFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/BloomFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/BloomFilter.java Fri Aug  3 15:39:43 2007
@@ -89,6 +89,7 @@
     vector = new boolean[this.vectorSize];
   }//end constructor
 
+  /** {@inheritDoc} */
   @Override
   public void add(Key key) {
     if(key == null) {
@@ -103,6 +104,7 @@
     }
   }//end add()
 
+  /** {@inheritDoc} */
   @Override
   public void and(Filter filter){
     if(filter == null
@@ -119,6 +121,7 @@
     }
   }//end and()
 
+  /** {@inheritDoc} */
   @Override
   public boolean membershipTest(Key key){
     if(key == null) {
@@ -135,6 +138,7 @@
     return true;
   }//end memberhsipTest()
 
+  /** {@inheritDoc} */
   @Override
   public void not(){
     for(int i = 0; i < vectorSize; i++) {
@@ -142,6 +146,7 @@
     }
   }//end not()
 
+  /** {@inheritDoc} */
   @Override
   public void or(Filter filter){
     if(filter == null
@@ -158,6 +163,7 @@
     }
   }//end or()
 
+  /** {@inheritDoc} */
   @Override
   public void xor(Filter filter){
     if(filter == null
@@ -175,7 +181,7 @@
     }
   }//and xor()
 
-  /** Returns a String representation of <i>this</i> Bloom filter. */
+  /** {@inheritDoc} */
   @Override
   public String toString(){
     StringBuilder res = new StringBuilder();
@@ -186,7 +192,7 @@
     return res.toString();
   }//end toString()
 
-  /** Returns a shallow copy of <i>this</i> Bloom filter. */
+  /** {@inheritDoc} */
   @Override
   public Object clone(){
     BloomFilter bf = new BloomFilter(vectorSize, nbHash);
@@ -194,11 +200,13 @@
     return bf;
   }//end clone()
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = super.hashCode();
@@ -210,6 +218,7 @@
 
   // Writable
 
+  /** {@inheritDoc} */
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
@@ -218,6 +227,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   @Override
   public void readFields(DataInput in) throws IOException {
     super.readFields(in);
@@ -229,6 +239,7 @@
 
   // Comparable
   
+  /** {@inheritDoc} */
   @Override
   public int compareTo(Object o) {
     int result = super.compareTo(o);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/CountingBloomFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/CountingBloomFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/CountingBloomFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/CountingBloomFilter.java Fri Aug  3 15:39:43 2007
@@ -84,6 +84,7 @@
     vector = new byte[vectorSize];
   }//end constructor
 
+  /** {@inheritDoc} */
   @Override
   public void add(Key key) {
     if(key == null) {
@@ -122,6 +123,7 @@
     }
   }//end delete
 
+  /** {@inheritDoc} */
   @Override
   public void and(Filter filter){
     if(filter == null
@@ -137,6 +139,7 @@
     }
   }//end and()
 
+  /** {@inheritDoc} */
   @Override
   public boolean membershipTest(Key key){
     if(key == null) {
@@ -155,12 +158,14 @@
     return true;
   }//end membershipTest()
 
+  /** {@inheritDoc} */
   @Override
   public void not(){
     throw new UnsupportedOperationException("not() is undefined for "
         + this.getClass().getName());
   }//end not()
 
+  /** {@inheritDoc} */
   @Override
   public void or(Filter filter){
     if(filter == null
@@ -177,6 +182,7 @@
     }
   }//end or()
 
+  /** {@inheritDoc} */
   @Override
   @SuppressWarnings("unused")
   public void xor(Filter filter){
@@ -184,6 +190,7 @@
         + this.getClass().getName());
   }//end xor()
 
+  /** {@inheritDoc} */
   @Override
   public String toString(){
     StringBuilder res = new StringBuilder();
@@ -198,6 +205,7 @@
     return res.toString();
   }//end toString()
 
+  /** {@inheritDoc} */
   @Override
   public Object clone(){
     CountingBloomFilter cbf = new CountingBloomFilter(vectorSize, nbHash);
@@ -205,11 +213,13 @@
     return cbf;
   }//end clone()
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = super.hashCode();
@@ -221,6 +231,7 @@
 
   // Writable
 
+  /** {@inheritDoc} */
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
@@ -229,6 +240,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   @Override
   public void readFields(DataInput in) throws IOException {
     super.readFields(in);
@@ -240,6 +252,7 @@
 
   // Comparable
   
+  /** {@inheritDoc} */
   @Override
   public int compareTo(Object o) {
     int result = super.compareTo(o);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/DynamicBloomFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/DynamicBloomFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/DynamicBloomFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/DynamicBloomFilter.java Fri Aug  3 15:39:43 2007
@@ -118,6 +118,7 @@
     matrix[0] = new BloomFilter(this.vectorSize, this.nbHash);
   }//end constructor
 
+  /** {@inheritDoc} */
   @Override
   public void add(Key key){
     if(key == null) {
@@ -137,6 +138,7 @@
     currentNbRecord++;
   }//end add()
 
+  /** {@inheritDoc} */
   @Override
   public void and(Filter filter) {
     if(filter == null
@@ -157,6 +159,7 @@
     }
   }//end and()
 
+  /** {@inheritDoc} */
   @Override
   public boolean membershipTest(Key key){
     if(key == null) {
@@ -172,6 +175,7 @@
     return false;
   }//end membershipTest()
 
+  /** {@inheritDoc} */
   @Override
   public void not(){
     for(int i = 0; i < matrix.length; i++) {
@@ -179,6 +183,7 @@
     }
   }//end not()
 
+  /** {@inheritDoc} */
   @Override
   public void or(Filter filter){
     if(filter == null
@@ -198,6 +203,7 @@
     }
   }//end or()
 
+  /** {@inheritDoc} */
   @Override
   public void xor(Filter filter){
     if(filter == null
@@ -217,6 +223,7 @@
     }
   }//end xor()
 
+  /** {@inheritDoc} */
   @Override
   public String toString(){
     StringBuilder res = new StringBuilder();
@@ -228,6 +235,7 @@
     return res.toString();
   }//end toString()
 
+  /** {@inheritDoc} */
   @Override
   public Object clone(){
     DynamicBloomFilter dbf = new DynamicBloomFilter(vectorSize, nbHash, nr);
@@ -239,11 +247,13 @@
     return dbf;
   }//end clone()
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = super.hashCode();
@@ -255,6 +265,7 @@
 
   // Writable
 
+  /** {@inheritDoc} */
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
@@ -263,6 +274,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   @Override
   public void readFields(DataInput in) throws IOException {
     super.readFields(in);
@@ -274,6 +286,7 @@
 
   // Comparable
   
+  /** {@inheritDoc} */
   @Override
   public int compareTo(Object o) {
     int result = super.compareTo(o);

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Filter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Filter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Filter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Filter.java Fri Aug  3 15:39:43 2007
@@ -182,6 +182,7 @@
     }
   }//end add()
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = Integer.valueOf(this.nbHash).hashCode();
@@ -191,17 +192,13 @@
 
   // Writable interface
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeInt(this.nbHash);
     out.writeInt(this.vectorSize);
   }
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.nbHash = in.readInt();
     this.vectorSize = in.readInt();
@@ -210,9 +207,7 @@
   
   // Comparable interface
   
-  /* (non-Javadoc)
-   * @see java.lang.Comparable#compareTo(java.lang.Object)
-   */
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     Filter other = (Filter)o;
     int result = this.vectorSize - other.vectorSize;

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Key.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Key.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Key.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/Key.java Fri Aug  3 15:39:43 2007
@@ -121,11 +121,13 @@
     this.weight++;
   }//end incrementWeight()
 
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = 0;
@@ -138,18 +140,14 @@
 
   // Writable
 
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#write(java.io.DataOutput)
-   */
+  /** {@inheritDoc} */
   public void write(DataOutput out) throws IOException {
     out.writeInt(bytes.length);
     out.write(bytes);
     out.writeDouble(weight);
   }
   
-  /* (non-Javadoc)
-   * @see org.apache.hadoop.io.Writable#readFields(java.io.DataInput)
-   */
+  /** {@inheritDoc} */
   public void readFields(DataInput in) throws IOException {
     this.bytes = new byte[in.readInt()];
     in.readFully(this.bytes);
@@ -158,9 +156,7 @@
   
   // Comparable
   
-  /* (non-Javadoc)
-   * @see java.lang.Comparable#compareTo(java.lang.Object)
-   */
+  /** {@inheritDoc} */
   public int compareTo(Object o) {
     Key other = (Key)o;
 

Modified: lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/RetouchedBloomFilter.java
URL: http://svn.apache.org/viewvc/lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/RetouchedBloomFilter.java?view=diff&rev=562608&r1=562607&r2=562608
==============================================================================
--- lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/RetouchedBloomFilter.java (original)
+++ lucene/hadoop/trunk/src/contrib/hbase/src/java/org/onelab/filter/RetouchedBloomFilter.java Fri Aug  3 15:39:43 2007
@@ -105,6 +105,7 @@
     createVector();
   }//end constructor
 
+  /** {@inheritDoc} */
   @Override
   public void add(Key key){
     if(key == null) {
@@ -392,11 +393,13 @@
     }//end for -i
   }//end createVector()
   
+  /** {@inheritDoc} */
   @Override
   public boolean equals(Object o) {
     return this.compareTo(o) == 0;
   }
   
+  /** {@inheritDoc} */
   @Override
   public int hashCode() {
     int result = super.hashCode();
@@ -414,6 +417,7 @@
 
   // Writable
 
+  /** {@inheritDoc} */
   @Override
   public void write(DataOutput out) throws IOException {
     super.write(out);
@@ -436,6 +440,7 @@
     }
   }
 
+  /** {@inheritDoc} */
   @Override
   public void readFields(DataInput in) throws IOException {
     super.readFields(in);
@@ -465,6 +470,7 @@
 
   // Comparable
   
+  /** {@inheritDoc} */
   @Override
   public int compareTo(Object o) {
     int result = super.compareTo(o);



Mime
View raw message