hadoop-hdfs-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From jgho...@apache.org
Subject svn commit: r983327 - in /hadoop/hdfs/trunk: ./ src/java/org/apache/hadoop/hdfs/ src/java/org/apache/hadoop/hdfs/server/datanode/ src/java/org/apache/hadoop/hdfs/server/namenode/
Date Sun, 08 Aug 2010 00:18:20 GMT
Author: jghoman
Date: Sun Aug  8 00:18:20 2010
New Revision: 983327

URL: http://svn.apache.org/viewvc?rev=983327&view=rev
Log:
HDFS-1297. Fixed some comments. Contributed by Jeff Ames.

Modified:
    hadoop/hdfs/trunk/CHANGES.txt
    hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/BlockReader.java
    hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java
    hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/Replica.java
    hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/DatanodeDescriptor.java
    hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java

Modified: hadoop/hdfs/trunk/CHANGES.txt
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/CHANGES.txt?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/CHANGES.txt (original)
+++ hadoop/hdfs/trunk/CHANGES.txt Sun Aug  8 00:18:20 2010
@@ -100,6 +100,8 @@ Trunk (unreleased changes)
     HDFS-1130. Adds dfs.cluster.administrator ACL configuration that can
     be used to control who can view the default hdfs servlets. (ddas)
 
+    HDFS-1297. Fix some comments. (Jeff Ames via jghoman)
+
   OPTIMIZATIONS
 
     HDFS-1140. Speedup INode.getPathComponents. (Dmytro Molkov via shv)

Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/BlockReader.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/BlockReader.java?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/BlockReader.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/BlockReader.java Sun Aug  8 00:18:20
2010
@@ -42,7 +42,7 @@ import org.apache.hadoop.net.NetUtils;
 import org.apache.hadoop.security.token.Token;
 import org.apache.hadoop.util.DataChecksum;
 
-/** This is a wrapper around connection to datadone
+/** This is a wrapper around connection to datanode
  * and understands checksum, offset etc
  */
 @InterfaceAudience.Private

Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java Sun
Aug  8 00:18:20 2010
@@ -196,7 +196,6 @@ class BlockSender implements java.io.Clo
         LOG.warn(datanode.dnRegistration + ":sendBlock() : " + msg);
         throw new IOException(msg);
       }
-
       
       offset = (startOffset - (startOffset % bytesPerChecksum));
       if (length >= 0) {
@@ -218,7 +217,7 @@ class BlockSender implements java.io.Clo
       // seek to the right offsets
       if (offset > 0) {
         long checksumSkip = (offset / bytesPerChecksum) * checksumSize;
-        // note blockInStream is  seeked when created below
+        // note blockInStream is seeked when created below
         if (checksumSkip > 0) {
           // Should we use seek() for checksum file as well?
           IOUtils.skipFully(checksumIn, checksumSkip);

Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/Replica.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/Replica.java?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/Replica.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/datanode/Replica.java Sun Aug
 8 00:18:20 2010
@@ -21,14 +21,14 @@ import org.apache.hadoop.classification.
 import org.apache.hadoop.hdfs.server.common.HdfsConstants.ReplicaState;
 
 /** 
- * This represents block replicas which stored in DataNode.
+ * This represents block replicas which are stored in DataNode.
  */
 @InterfaceAudience.Private
 public interface Replica {
-  /** get block ID  */
+  /** Get the block ID  */
   public long getBlockId();
 
-  /** get generation stamp */
+  /** Get the generation stamp */
   public long getGenerationStamp();
 
   /**

Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/DatanodeDescriptor.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/DatanodeDescriptor.java?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/DatanodeDescriptor.java
(original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/DatanodeDescriptor.java
Sun Aug  8 00:18:20 2010
@@ -38,15 +38,13 @@ import org.apache.hadoop.hdfs.Deprecated
 import org.apache.hadoop.io.WritableUtils;
 
 /**************************************************
- * DatanodeDescriptor tracks stats on a given DataNode,
- * such as available storage capacity, last update time, etc.,
- * and maintains a set of blocks stored on the datanode. 
+ * DatanodeDescriptor tracks stats on a given DataNode, such as
+ * available storage capacity, last update time, etc., and maintains a
+ * set of blocks stored on the datanode.
  *
- * This data structure is a data structure that is internal
- * to the namenode. It is *not* sent over-the-wire to the Client
- * or the Datnodes. Neither is it stored persistently in the
- * fsImage.
-
+ * This data structure is internal to the namenode. It is *not* sent
+ * over-the-wire to the Client or the Datanodes. Neither is it stored
+ * persistently in the fsImage.
  **************************************************/
 @InterfaceAudience.Private
 public class DatanodeDescriptor extends DatanodeInfo {

Modified: hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java
URL: http://svn.apache.org/viewvc/hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java?rev=983327&r1=983326&r2=983327&view=diff
==============================================================================
--- hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java (original)
+++ hadoop/hdfs/trunk/src/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java Sun Aug
 8 00:18:20 2010
@@ -641,7 +641,7 @@ public class FSEditLog {
           numOpOther++;
           FSImage.DatanodeImage nodeimage = new FSImage.DatanodeImage();
           nodeimage.readFields(in);
-          //Datnodes are not persistent any more.
+          //Datanodes are not persistent any more.
           break;
         }
         case OP_DATANODE_REMOVE: {



Mime
View raw message