hadoop-common-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From omal...@apache.org
Subject svn commit: r673857 [5/6] - in /hadoop/core/trunk: ./ bin/ conf/ docs/ src/contrib/eclipse-plugin/src/java/org/apache/hadoop/eclipse/ src/contrib/eclipse-plugin/src/java/org/apache/hadoop/eclipse/dfs/ src/contrib/index/src/java/org/apache/hadoop/contri...
Date Thu, 03 Jul 2008 22:55:18 GMT
Copied: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamenodeProtocol.java (from r673837, hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamenodeProtocol.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamenodeProtocol.java?p2=hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamenodeProtocol.java&p1=hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamenodeProtocol.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamenodeProtocol.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamenodeProtocol.java Thu Jul  3 15:55:06 2008
@@ -16,17 +16,19 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs.server.protocol;
 
 import java.io.IOException;
 
+import org.apache.hadoop.hdfs.protocol.BlocksWithLocations;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
 import org.apache.hadoop.ipc.VersionedProtocol;
 
 /*****************************************************************************
  * Protocol that a secondary NameNode uses to communicate with the NameNode.
  * It's used to get part of the name node state
  *****************************************************************************/
-interface NamenodeProtocol extends VersionedProtocol {
+public interface NamenodeProtocol extends VersionedProtocol {
   /**
    * 1: changed the serialization in DatanodeInfo
    */
@@ -37,7 +39,7 @@
    * @param datanode  a data node
    * @param size      requested size
    * @return          a list of blocks & their locations
-   * @Exception RemoteException if size is less than or equal to 0 or
+   * @throws RemoteException if size is less than or equal to 0 or
                                    datanode does not exist
    */
   public BlocksWithLocations getBlocks(DatanodeInfo datanode, long size)

Copied: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamespaceInfo.java (from r673837, hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamespaceInfo.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamespaceInfo.java?p2=hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamespaceInfo.java&p1=hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamespaceInfo.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/NamespaceInfo.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/NamespaceInfo.java Thu Jul  3 15:55:06 2008
@@ -16,12 +16,15 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs.server.protocol;
 
 import java.io.DataInput;
 import java.io.DataOutput;
 import java.io.IOException;
 
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.server.common.Storage;
+import org.apache.hadoop.hdfs.server.common.StorageInfo;
 import org.apache.hadoop.io.UTF8;
 import org.apache.hadoop.io.Writable;
 import org.apache.hadoop.io.WritableFactories;
@@ -32,7 +35,7 @@
  * to a data-node handshake.
  * 
  */
-class NamespaceInfo extends StorageInfo implements Writable {
+public class NamespaceInfo extends StorageInfo implements Writable {
   String  buildVersion;
   int distributedUpgradeVersion;
 

Copied: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/UpgradeCommand.java (from r673837, hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/UpgradeCommand.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/UpgradeCommand.java?p2=hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/UpgradeCommand.java&p1=hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/UpgradeCommand.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/UpgradeCommand.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/server/protocol/UpgradeCommand.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs.server.protocol;
 
 import java.io.DataInput;
 import java.io.DataOutput;
@@ -35,31 +35,31 @@
  * The upgrade command contains version of the upgrade, which is verified 
  * on the receiving side and current status of the upgrade.
  */
-class UpgradeCommand extends DatanodeCommand {
+public class UpgradeCommand extends DatanodeCommand {
   final static int UC_ACTION_UNKNOWN = DatanodeProtocol.DNA_UNKNOWN;
-  final static int UC_ACTION_REPORT_STATUS = 100; // report upgrade status
-  final static int UC_ACTION_START_UPGRADE = 101; // start upgrade
+  public final static int UC_ACTION_REPORT_STATUS = 100; // report upgrade status
+  public final static int UC_ACTION_START_UPGRADE = 101; // start upgrade
 
   private int version;
   private short upgradeStatus;
 
-  UpgradeCommand() {
+  public UpgradeCommand() {
     super(UC_ACTION_UNKNOWN);
     this.version = 0;
     this.upgradeStatus = 0;
   }
 
-  UpgradeCommand(int action, int version, short status) {
+  public UpgradeCommand(int action, int version, short status) {
     super(action);
     this.version = version;
     this.upgradeStatus = status;
   }
 
-  int getVersion() {
+  public int getVersion() {
     return this.version;
   }
 
-  short getCurrentStatus() {
+  public short getCurrentStatus() {
     return this.upgradeStatus;
   }
 

Copied: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSAdmin.java (from r673837, hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSAdmin.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSAdmin.java?p2=hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSAdmin.java&p1=hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSAdmin.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSAdmin.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSAdmin.java Thu Jul  3 15:55:06 2008
@@ -15,14 +15,18 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs.tools;
 
 import java.io.IOException;
 import java.util.List;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.DistributedFileSystem.DiskStatus;
-import org.apache.hadoop.dfs.FSConstants.UpgradeAction;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
+import org.apache.hadoop.hdfs.DistributedFileSystem.DiskStatus;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.FSConstants.UpgradeAction;
+import org.apache.hadoop.hdfs.server.common.UpgradeStatusReport;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FsShell;
 import org.apache.hadoop.fs.Path;

Copied: hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSck.java (from r673837, hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSck.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSck.java?p2=hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSck.java&p1=hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSck.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/hdfs/org/apache/hadoop/dfs/DFSck.java (original)
+++ hadoop/core/trunk/src/hdfs/org/apache/hadoop/hdfs/tools/DFSck.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs.tools;
 
 import java.io.IOException;
 import java.io.InputStream;
@@ -40,11 +40,11 @@
  * <li>files with blocks that are completely missing from all datanodes.<br/>
  * In this case the tool can perform one of the following actions:
  *  <ul>
- *      <li>none ({@link NamenodeFsck#FIXING_NONE})</li>
+ *      <li>none ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_NONE})</li>
  *      <li>move corrupted files to /lost+found directory on DFS
- *      ({@link NamenodeFsck#FIXING_MOVE}). Remaining data blocks are saved as a
+ *      ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_MOVE}). Remaining data blocks are saved as a
  *      block chains, representing longest consecutive series of valid blocks.</li>
- *      <li>delete corrupted files ({@link NamenodeFsck#FIXING_DELETE})</li>
+ *      <li>delete corrupted files ({@link org.apache.hadoop.hdfs.server.namenode.NamenodeFsck#FIXING_DELETE})</li>
  *  </ul>
  *  </li>
  *  <li>detect files with under-replicated or over-replicated blocks</li>

Modified: hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/JobID.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/JobID.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/JobID.java (original)
+++ hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/JobID.java Thu Jul  3 15:55:06 2008
@@ -180,4 +180,4 @@
     return builder;
   }
   
-}
\ No newline at end of file
+}

Modified: hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/Task.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/Task.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/Task.java (original)
+++ hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/Task.java Thu Jul  3 15:55:06 2008
@@ -33,7 +33,7 @@
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configurable;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.DistributedFileSystem;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
 import org.apache.hadoop.fs.ContentSummary;
 import org.apache.hadoop.fs.FileStatus;
 import org.apache.hadoop.fs.FileSystem;

Modified: hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskAttemptID.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskAttemptID.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskAttemptID.java (original)
+++ hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskAttemptID.java Thu Jul  3 15:55:06 2008
@@ -209,4 +209,4 @@
     return builder;
   }
   
-}
\ No newline at end of file
+}

Modified: hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskID.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskID.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskID.java (original)
+++ hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/TaskID.java Thu Jul  3 15:55:06 2008
@@ -221,4 +221,4 @@
     return builder;
   }
   
-}
\ No newline at end of file
+}

Modified: hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/lib/aggregate/ValueAggregator.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/lib/aggregate/ValueAggregator.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/lib/aggregate/ValueAggregator.java (original)
+++ hadoop/core/trunk/src/mapred/org/apache/hadoop/mapred/lib/aggregate/ValueAggregator.java Thu Jul  3 15:55:06 2008
@@ -50,4 +50,4 @@
    */
   public ArrayList getCombinerOutput();
 
-}
\ No newline at end of file
+}

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/cli/TestCLI.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/cli/TestCLI.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/cli/TestCLI.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/cli/TestCLI.java Thu Jul  3 15:55:06 2008
@@ -36,9 +36,9 @@
 
 import org.apache.hadoop.cli.util.CommandExecutor;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.DataNode;
-import org.apache.hadoop.dfs.DistributedFileSystem;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.DistributedFileSystem;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FsShell;
 import org.apache.hadoop.fs.Path;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestCopyFiles.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestCopyFiles.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestCopyFiles.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestCopyFiles.java Thu Jul  3 15:55:06 2008
@@ -25,7 +25,7 @@
 import junit.framework.TestCase;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 import org.apache.hadoop.fs.permission.FsPermission;
 import org.apache.hadoop.mapred.JobConf;
 import org.apache.hadoop.mapred.MiniMRCluster;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestFileSystem.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestFileSystem.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestFileSystem.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestFileSystem.java Thu Jul  3 15:55:06 2008
@@ -31,7 +31,7 @@
 import org.apache.commons.logging.Log;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.conf.Configured;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 import org.apache.hadoop.fs.shell.CommandFormat;
 import org.apache.hadoop.io.LongWritable;
 import org.apache.hadoop.io.SequenceFile;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestGlobPaths.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestGlobPaths.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestGlobPaths.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestGlobPaths.java Thu Jul  3 15:55:06 2008
@@ -20,7 +20,7 @@
 import java.io.IOException;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 
 import junit.framework.TestCase;
 

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestHarFileSystem.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestHarFileSystem.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestHarFileSystem.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestHarFileSystem.java Thu Jul  3 15:55:06 2008
@@ -23,7 +23,7 @@
 import java.util.Iterator;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FsShell;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestUrlStreamHandler.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestUrlStreamHandler.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestUrlStreamHandler.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/TestUrlStreamHandler.java Thu Jul  3 15:55:06 2008
@@ -28,7 +28,7 @@
 import junit.framework.TestCase;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FsUrlStreamHandlerFactory;
 import org.apache.hadoop.fs.Path;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/ftp/TestFTPFileSystem.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/ftp/TestFTPFileSystem.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/ftp/TestFTPFileSystem.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/ftp/TestFTPFileSystem.java Thu Jul  3 15:55:06 2008
@@ -28,7 +28,7 @@
 import org.apache.ftpserver.usermanager.BaseUser;
 import org.apache.ftpserver.usermanager.WritePermission;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.DFSTestUtil;
+import org.apache.hadoop.hdfs.DFSTestUtil;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FileUtil;
 import org.apache.hadoop.fs.Path;

Modified: hadoop/core/trunk/src/test/org/apache/hadoop/fs/permission/TestFsPermission.java
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/fs/permission/TestFsPermission.java?rev=673857&r1=673856&r2=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/fs/permission/TestFsPermission.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/fs/permission/TestFsPermission.java Thu Jul  3 15:55:06 2008
@@ -67,4 +67,4 @@
       assertEquals(i, FsPermission.valueOf(b.toString()).toShort());
     }
   }
-}
\ No newline at end of file
+}

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/BenchmarkThroughput.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/BenchmarkThroughput.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/BenchmarkThroughput.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/BenchmarkThroughput.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/BenchmarkThroughput.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/BenchmarkThroughput.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/BenchmarkThroughput.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.FileInputStream;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFS.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFS.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFS.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFS.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFS.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFS.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFS.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import junit.framework.AssertionFailedError;
@@ -25,9 +25,14 @@
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FSInputStream;
 import org.apache.hadoop.fs.FileUtil;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
 import org.apache.hadoop.io.UTF8;
 import org.apache.hadoop.conf.Configuration;
 
+
 import java.io.File;
 import java.io.OutputStream;
 import java.net.InetSocketAddress;
@@ -71,7 +76,7 @@
  */
 public class ClusterTestDFS extends TestCase implements FSConstants {
   private static final Log LOG =
-    LogFactory.getLog("org.apache.hadoop.dfs.ClusterTestDFS");
+    LogFactory.getLog("org.apache.hadoop.hdfs.ClusterTestDFS");
 
   private static Configuration conf = new Configuration();
   private static int BUFFER_SIZE =

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFSNamespaceLogging.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFSNamespaceLogging.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFSNamespaceLogging.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFSNamespaceLogging.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFSNamespaceLogging.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/ClusterTestDFSNamespaceLogging.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/ClusterTestDFSNamespaceLogging.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import junit.framework.AssertionFailedError;
@@ -27,7 +27,9 @@
 import org.apache.hadoop.fs.FileUtil;
 import org.apache.hadoop.io.UTF8;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.NameNode;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
 
 import java.io.BufferedReader;
 import java.io.File;
@@ -44,7 +46,7 @@
  */
 public class ClusterTestDFSNamespaceLogging extends TestCase implements FSConstants {
   private static final Log LOG =
-    LogFactory.getLog("org.apache.hadoop.dfs.ClusterTestDFS");
+    LogFactory.getLog("org.apache.hadoop.hdfs.ClusterTestDFS");
 
   private static Configuration conf = new Configuration();
 

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/CreateEditsLog.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/CreateEditsLog.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/CreateEditsLog.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/CreateEditsLog.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/CreateEditsLog.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/CreateEditsLog.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/CreateEditsLog.java Thu Jul  3 15:55:06 2008
@@ -15,13 +15,20 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
 
 
-import org.apache.hadoop.dfs.BlocksMap.BlockInfo;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.server.common.GenerationStamp;
+import org.apache.hadoop.hdfs.server.common.Storage;
+import org.apache.hadoop.hdfs.server.namenode.FSEditLog;
+import org.apache.hadoop.hdfs.server.namenode.FSImage;
+import org.apache.hadoop.hdfs.server.namenode.INodeDirectory;
+import org.apache.hadoop.hdfs.server.namenode.INodeFileUnderConstruction;
+import org.apache.hadoop.hdfs.server.namenode.BlocksMap.BlockInfo;
 import org.apache.hadoop.fs.permission.FsPermission;
 import org.apache.hadoop.fs.permission.PermissionStatus;
 
@@ -72,7 +79,7 @@
 
     for (int iF = 1; iF <= numFiles; iF++) {
       for (int iB = 0; iB < blocksPerFile; ++iB) {
-         blocks[iB].blkid = currentBlockId++;
+         blocks[iB].setBlockId(currentBlockId++);
       }
 
       try {

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DFSTestUtil.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DFSTestUtil.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DFSTestUtil.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DFSTestUtil.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DFSTestUtil.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DFSTestUtil.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DFSTestUtil.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.BufferedReader;
 import java.io.File;
@@ -25,7 +25,8 @@
 import java.util.Random;
 import junit.framework.TestCase;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.DFSClient.DFSDataInputStream;
+import org.apache.hadoop.hdfs.DFSClient.DFSDataInputStream;
+import org.apache.hadoop.hdfs.protocol.Block;
 import org.apache.hadoop.fs.FSDataInputStream;
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
@@ -117,7 +118,7 @@
     }
   }
   
-  static void createFile(FileSystem fs, Path fileName, long fileLen, 
+  public static void createFile(FileSystem fs, Path fileName, long fileLen, 
       short replFactor, long seed) throws IOException {
     if (!fs.mkdirs(fileName.getParent())) {
       throw new IOException("Mkdirs failed to create " + 
@@ -176,7 +177,7 @@
   // waits for the replication factor of all files to reach the
   // specified target
   //
-  void waitReplication(FileSystem fs, String topdir, short value) 
+  public void waitReplication(FileSystem fs, String topdir, short value) 
                                               throws IOException {
     Path root = new Path(topdir);
 
@@ -200,7 +201,7 @@
   }
   
   /** wait for the file's replication to be done */
-  static void waitReplication(FileSystem fs, Path fileName, 
+  public static void waitReplication(FileSystem fs, Path fileName, 
       short replFactor)  throws IOException {
     boolean good;
     do {
@@ -224,13 +225,13 @@
   }
   
   /** delete directory and everything underneath it.*/
-  void cleanup(FileSystem fs, String topdir) throws IOException {
+  public void cleanup(FileSystem fs, String topdir) throws IOException {
     Path root = new Path(topdir);
     fs.delete(root, true);
     files = null;
   }
   
-  static Block getFirstBlock(FileSystem fs, Path path) throws IOException {
+  public static Block getFirstBlock(FileSystem fs, Path path) throws IOException {
     DFSDataInputStream in = 
       (DFSDataInputStream) ((DistributedFileSystem)fs).open(path);
     in.readByte();

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DataNodeCluster.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DataNodeCluster.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DataNodeCluster.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DataNodeCluster.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DataNodeCluster.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/DataNodeCluster.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/DataNodeCluster.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 import java.net.UnknownHostException;
@@ -24,7 +24,9 @@
 import java.util.Random;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.net.DNS;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/MiniDFSCluster.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/MiniDFSCluster.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/MiniDFSCluster.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/MiniDFSCluster.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/MiniDFSCluster.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/MiniDFSCluster.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/MiniDFSCluster.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -31,8 +31,16 @@
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.net.*;
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.FSDatasetInterface;
+import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
+import org.apache.hadoop.hdfs.tools.DFSAdmin;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FileUtil;
 import org.apache.hadoop.security.*;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBench.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBench.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBench.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBench.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBench.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBench.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBench.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 import java.util.Date;
@@ -79,7 +79,7 @@
 
 public class NNBench {
   private static final Log LOG = LogFactory.getLog(
-          "org.apache.hadoop.dfs.NNBench");
+          "org.apache.hadoop.hdfs.NNBench");
   
   protected static String CONTROL_DIR_NAME = "control";
   protected static String OUTPUT_DIR_NAME = "output";

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBenchWithoutMR.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBenchWithoutMR.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBenchWithoutMR.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBenchWithoutMR.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBenchWithoutMR.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNBenchWithoutMR.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNBenchWithoutMR.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 import java.util.Date;
@@ -46,7 +46,7 @@
 public class NNBenchWithoutMR {
   
   private static final Log LOG = LogFactory.getLog(
-                                            "org.apache.hadoop.dfs.NNBench");
+                                            "org.apache.hadoop.hdfs.NNBench");
   
   // variable initialzed from command line arguments
   private static long startTime = 0;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNThroughputBenchmark.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNThroughputBenchmark.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNThroughputBenchmark.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNThroughputBenchmark.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNThroughputBenchmark.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/NNThroughputBenchmark.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/NNThroughputBenchmark.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.FileOutputStream;
@@ -31,6 +31,20 @@
 import org.apache.commons.logging.impl.Log4JLogger;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.permission.FsPermission;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.BlockListAsLongs;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.DataStorage;
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
+import org.apache.hadoop.hdfs.server.protocol.BlockCommand;
+import org.apache.hadoop.hdfs.server.protocol.DatanodeCommand;
+import org.apache.hadoop.hdfs.server.protocol.DatanodeProtocol;
+import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration;
+import org.apache.hadoop.hdfs.server.protocol.NamespaceInfo;
 import org.apache.hadoop.net.DNS;
 import org.apache.hadoop.net.NetworkTopology;
 import org.apache.hadoop.util.StringUtils;
@@ -63,7 +77,7 @@
  * the specified number of threads and outputs the resulting stats.
  */
 public class NNThroughputBenchmark implements FSConstants {
-  private static final Log LOG = LogFactory.getLog("org.apache.hadoop.dfs.NNThroughputBenchmark");
+  private static final Log LOG = LogFactory.getLog("org.apache.hadoop.hdfs.NNThroughputBenchmark");
   private static final int BLOCK_SIZE = 16;
 
   static Configuration config;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestAbandonBlock.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestAbandonBlock.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestAbandonBlock.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestAbandonBlock.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestAbandonBlock.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestAbandonBlock.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestAbandonBlock.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 
@@ -23,6 +23,8 @@
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.*;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
 import org.apache.hadoop.util.StringUtils;
 
 public class TestAbandonBlock extends junit.framework.TestCase {

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestBlockReplacement.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestBlockReplacement.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestBlockReplacement.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestBlockReplacement.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestBlockReplacement.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestBlockReplacement.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestBlockReplacement.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.DataInputStream;
 import java.io.DataOutputStream;
@@ -30,19 +30,27 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.server.common.Util;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.net.NetUtils;
 
+
 import junit.framework.TestCase;
 /**
  * This class tests if block replacement request to data nodes work correctly.
  */
 public class TestBlockReplacement extends TestCase {
   private static final Log LOG = LogFactory.getLog(
-  "org.apache.hadoop.dfs.TestBlockReplacement");
+  "org.apache.hadoop.hdfs.TestBlockReplacement");
 
   MiniDFSCluster cluster;
   public void testThrottler() throws IOException {
@@ -51,7 +59,7 @@
     long bandwidthPerSec = 1024*1024L;
     final long TOTAL_BYTES =6*bandwidthPerSec; 
     long bytesToSend = TOTAL_BYTES; 
-    long start = FSNamesystem.now();
+    long start = Util.now();
     DataNode.Throttler throttler = new DataNode.Throttler(bandwidthPerSec);
     long totalBytes = 0L;
     long bytesSent = 1024*512L; // 0.5MB
@@ -64,7 +72,7 @@
       Thread.sleep(1000);
     } catch (InterruptedException ignored) {}
     throttler.throttle(bytesToSend);
-    long end = FSNamesystem.now();
+    long end = Util.now();
     assertTrue(totalBytes*1000/(end-start)<=bandwidthPerSec);
   }
   

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestCrcCorruption.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestCrcCorruption.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestCrcCorruption.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestCrcCorruption.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestCrcCorruption.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestCrcCorruption.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestCrcCorruption.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.*;
 import java.nio.channels.FileChannel;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSFinalize.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSFinalize.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSFinalize.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSFinalize.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSFinalize.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSFinalize.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSFinalize.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -23,9 +23,9 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.NAME_NODE;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.DATA_NODE;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.NAME_NODE;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.DATA_NODE;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
 
 /**
  * This test ensures the appropriate response from the system when 
@@ -34,7 +34,7 @@
 public class TestDFSFinalize extends TestCase {
  
   private static final Log LOG = LogFactory.getLog(
-                                                   "org.apache.hadoop.dfs.TestDFSFinalize");
+                                                   "org.apache.hadoop.hdfs.TestDFSFinalize");
   private Configuration conf;
   private int testCounter = 0;
   private MiniDFSCluster cluster = null;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSMkdirs.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSMkdirs.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSMkdirs.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSMkdirs.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSMkdirs.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSMkdirs.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSMkdirs.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSPermission.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSPermission.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSPermission.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSPermission.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSPermission.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSPermission.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSPermission.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 import java.util.Random;
@@ -24,7 +24,8 @@
 
 import org.apache.commons.logging.*;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.MiniDFSCluster;
+import org.apache.hadoop.hdfs.server.common.Util;
 import org.apache.hadoop.fs.*;
 import org.apache.hadoop.fs.permission.*;
 import org.apache.hadoop.security.UnixUserGroupInformation;
@@ -70,7 +71,7 @@
   static {
     try {
       // Initiate the random number generator and logging the seed
-      long seed = FSNamesystem.now();
+      long seed = Util.now();
       r = new Random(seed);
       LOG.info("Random number generator uses seed " + seed);
       LOG.info("NUM_TEST_PERMISSIONS=" + NUM_TEST_PERMISSIONS);
@@ -957,4 +958,4 @@
   private void checkNoPermissionDeny(IOException e) {
     assertFalse(e instanceof AccessControlException);
   }
-}
\ No newline at end of file
+}

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRename.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRename.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRename.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRename.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRename.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRename.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRename.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.*;
 

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRollback.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRollback.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRollback.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRollback.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRollback.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSRollback.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSRollback.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -23,10 +23,11 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.NodeType;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.NAME_NODE;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.DATA_NODE;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.FSConstants.NodeType;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.NAME_NODE;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.DATA_NODE;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.server.common.StorageInfo;
 import org.apache.hadoop.fs.FileUtil;
 
 /**
@@ -37,7 +38,7 @@
 public class TestDFSRollback extends TestCase {
  
   private static final Log LOG = LogFactory.getLog(
-                                                   "org.apache.hadoop.dfs.TestDFSRollback");
+                                                   "org.apache.hadoop.hdfs.TestDFSRollback");
   private Configuration conf;
   private int testCounter = 0;
   private MiniDFSCluster cluster = null;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShell.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShell.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShell.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShell.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShell.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShell.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShell.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -27,6 +27,9 @@
 import org.apache.hadoop.fs.*;
 import org.apache.hadoop.fs.permission.FsPermission;
 import org.apache.hadoop.fs.shell.*;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.FSDataset;
 import org.apache.hadoop.io.IOUtils;
 import org.apache.hadoop.security.UnixUserGroupInformation;
 import org.apache.hadoop.util.StringUtils;
@@ -563,7 +566,7 @@
     /* This tests some properties of ChecksumFileSystem as well.
      * Make sure that we create ChecksumDFS */
     conf.set("fs.hdfs.impl",
-             "org.apache.hadoop.dfs.ChecksumDistributedFileSystem");
+             "org.apache.hadoop.hdfs.ChecksumDistributedFileSystem");
     MiniDFSCluster cluster = new MiniDFSCluster(conf, 2, true, null);
     FileSystem fs = cluster.getFileSystem();
     assertTrue("Not a HDFS: "+fs.getUri(),
@@ -841,7 +844,7 @@
     /* This tests some properties of ChecksumFileSystem as well.
      * Make sure that we create ChecksumDFS */
     conf.set("fs.hdfs.impl",
-             "org.apache.hadoop.dfs.ChecksumDistributedFileSystem");
+             "org.apache.hadoop.hdfs.ChecksumDistributedFileSystem");
     MiniDFSCluster cluster = new MiniDFSCluster(conf, 2, true, null);
     FileSystem fs = cluster.getFileSystem();
     assertTrue("Not a HDFS: "+fs.getUri(),

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShellGenericOptions.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShellGenericOptions.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShellGenericOptions.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShellGenericOptions.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShellGenericOptions.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSShellGenericOptions.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSShellGenericOptions.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.FileNotFoundException;
@@ -28,6 +28,7 @@
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FsShell;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
 import org.apache.hadoop.net.NetUtils;
 import org.apache.hadoop.util.ToolRunner;
 

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStartupVersions.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStartupVersions.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStartupVersions.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStartupVersions.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStartupVersions.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStartupVersions.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStartupVersions.java Thu Jul  3 15:55:06 2008
@@ -15,17 +15,20 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import junit.framework.TestCase;
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.NodeType;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.NAME_NODE;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.DATA_NODE;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.FSConstants.NodeType;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.NAME_NODE;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.DATA_NODE;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.server.common.Storage;
+import org.apache.hadoop.hdfs.server.common.StorageInfo;
 import org.apache.hadoop.fs.Path;
 
 /**
@@ -35,7 +38,7 @@
 public class TestDFSStartupVersions extends TestCase {
   
   private static final Log LOG = LogFactory.getLog(
-                                                   "org.apache.hadoop.dfs.TestDFSStartupVersions");
+                                                   "org.apache.hadoop.hdfs.TestDFSStartupVersions");
   private static Path TEST_ROOT_DIR = new Path(
                                                System.getProperty("test.build.data","/tmp").toString().replace(' ', '+'));
   private MiniDFSCluster cluster = null;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStorageStateRecovery.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStorageStateRecovery.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSStorageStateRecovery.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSStorageStateRecovery.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -23,10 +23,10 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.NodeType;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.NAME_NODE;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.DATA_NODE;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.FSConstants.NodeType;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.NAME_NODE;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.DATA_NODE;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
 
 /**
 * This test ensures the appropriate response (successful or failure) from
@@ -36,7 +36,7 @@
 public class TestDFSStorageStateRecovery extends TestCase {
  
   private static final Log LOG = LogFactory.getLog(
-                                                   "org.apache.hadoop.dfs.TestDFSStorageStateRecovery");
+                                                   "org.apache.hadoop.hdfs.TestDFSStorageStateRecovery");
   private Configuration conf = null;
   private int testCounter = 0;
   private MiniDFSCluster cluster = null;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgrade.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgrade.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgrade.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgrade.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgrade.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgrade.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgrade.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -23,10 +23,12 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.NodeType;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.NAME_NODE;
-import static org.apache.hadoop.dfs.FSConstants.NodeType.DATA_NODE;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.FSConstants.NodeType;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.NAME_NODE;
+import static org.apache.hadoop.hdfs.protocol.FSConstants.NodeType.DATA_NODE;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.server.common.Storage;
+import org.apache.hadoop.hdfs.server.common.StorageInfo;
 import org.apache.hadoop.fs.FileUtil;
 
 /**
@@ -37,7 +39,7 @@
 public class TestDFSUpgrade extends TestCase {
  
   private static final Log LOG = LogFactory.getLog(
-                                                   "org.apache.hadoop.dfs.TestDFSUpgrade");
+                                                   "org.apache.hadoop.hdfs.TestDFSUpgrade");
   private Configuration conf;
   private int testCounter = 0;
   private MiniDFSCluster cluster = null;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgradeFromImage.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgradeFromImage.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDFSUpgradeFromImage.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -29,7 +29,9 @@
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FSInputStream;
 import org.apache.hadoop.fs.FileUtil;
-import org.apache.hadoop.dfs.FSConstants.StartupOption;
+import org.apache.hadoop.hdfs.protocol.DFSFileInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.FSConstants.StartupOption;
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 
@@ -45,9 +47,9 @@
 public class TestDFSUpgradeFromImage extends TestCase {
   
   private static final Log LOG = LogFactory.getLog(
-                    "org.apache.hadoop.dfs.TestDFSUpgradeFromImage");
+                    "org.apache.hadoop.hdfs.TestDFSUpgradeFromImage");
   
-  int numDataNodes = 4;
+  public int numDataNodes = 4;
   
   private static class ReferenceFileInfo {
     String path;
@@ -63,10 +65,10 @@
     unpackStorage();
   }
 
-  void unpackStorage() throws IOException {
-    String tarFile = System.getProperty("test.cache.data") + 
+  public void unpackStorage() throws IOException {
+    String tarFile = System.getProperty("test.cache.data", "build/test/cache") +
                      "/hadoop-14-dfs-dir.tgz";
-    String dataDir = System.getProperty("test.build.data");
+    String dataDir = System.getProperty("test.build.data", "build/test/data");
     File dfsDir = new File(dataDir, "dfs");
     if ( dfsDir.exists() && !FileUtil.fullyDelete(dfsDir) ) {
       throw new IOException("Could not delete dfs directory '" + dfsDir + "'");
@@ -75,7 +77,7 @@
     //Now read the reference info
     
     BufferedReader reader = new BufferedReader( 
-                        new FileReader(System.getProperty("test.cache.data") +
+                        new FileReader(System.getProperty("test.cache.data", "build/test/cache") +
                                        "/hadoop-dfs-dir.txt"));
     String line;
     while ( (line = reader.readLine()) != null ) {
@@ -176,6 +178,9 @@
     MiniDFSCluster cluster = null;
     try {
       Configuration conf = new Configuration();
+      if (System.getProperty("test.build.data") == null) { // to allow test to be run outside of Ant
+        System.setProperty("test.build.data", "build/test/data");
+      }
       conf.setInt("dfs.datanode.scan.period.hours", -1); // block scanning off
       cluster = new MiniDFSCluster(0, conf, numDataNodes, false, true,
                                    StartupOption.UPGRADE, null);

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDataTransferProtocol.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDataTransferProtocol.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDataTransferProtocol.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDataTransferProtocol.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDataTransferProtocol.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDataTransferProtocol.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDataTransferProtocol.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -30,9 +30,13 @@
 import org.apache.hadoop.io.IOUtils;
 import org.apache.hadoop.io.Text;
 import org.apache.hadoop.net.NetUtils;
+import org.apache.hadoop.util.DataChecksum;
 import org.apache.hadoop.fs.Path;
-import org.apache.hadoop.dfs.DFSClient.DFSDataInputStream;
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.DFSClient.DFSDataInputStream;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeID;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 
@@ -43,7 +47,7 @@
 public class TestDataTransferProtocol extends TestCase {
   
   private static final Log LOG = LogFactory.getLog(
-                    "org.apache.hadoop.dfs.TestDataTransferProtocol");
+                    "org.apache.hadoop.hdfs.TestDataTransferProtocol");
   
   DatanodeID datanode;
   InetSocketAddress dnAddr;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeBlockScanner.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeBlockScanner.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeBlockScanner.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeBlockScanner.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeBlockScanner.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeBlockScanner.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeBlockScanner.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.ByteArrayOutputStream;
 import java.io.IOException;
@@ -32,7 +32,10 @@
 import org.apache.commons.logging.Log;
 import org.apache.commons.logging.LogFactory;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
 import org.apache.hadoop.io.IOUtils;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeDeath.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeDeath.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeDeath.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeDeath.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeDeath.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeDeath.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeDeath.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -30,6 +30,8 @@
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FSDataInputStream;
 import org.apache.hadoop.fs.BlockLocation;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
 
 /**
  * This class tests that a file need not be closed before its

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeReport.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeReport.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeReport.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeReport.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeReport.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDatanodeReport.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDatanodeReport.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.net.InetSocketAddress;
 import java.util.ArrayList;
@@ -23,7 +23,9 @@
 import junit.framework.TestCase;
 import org.apache.hadoop.conf.Configuration;
 
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
 
 /**
  * This test ensures the all types of data node report work correctly.

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDecommission.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDecommission.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDecommission.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDecommission.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDecommission.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDecommission.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDecommission.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -26,7 +26,9 @@
 import java.net.*;
 import java.lang.InterruptedException;
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.FSConstants.DatanodeReportType;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.FSConstants.DatanodeReportType;
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDefaultNameNodePort.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDefaultNameNodePort.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDefaultNameNodePort.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDefaultNameNodePort.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDefaultNameNodePort.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDefaultNameNodePort.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDefaultNameNodePort.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -24,6 +24,7 @@
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
 
 /** Test NameNode port defaulting code. */
 public class TestDefaultNameNodePort extends TestCase {

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDistributedFileSystem.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDistributedFileSystem.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDistributedFileSystem.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDistributedFileSystem.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDistributedFileSystem.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestDistributedFileSystem.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestDistributedFileSystem.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.net.URI;
 

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSInputChecker.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSInputChecker.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSInputChecker.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSInputChecker.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSInputChecker.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSInputChecker.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSInputChecker.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -221,7 +221,7 @@
     conf.setLong("dfs.block.size", BLOCK_SIZE);
     conf.setInt("io.bytes.per.checksum", BYTES_PER_SUM);
     conf.set("fs.hdfs.impl",
-             "org.apache.hadoop.dfs.ChecksumDistributedFileSystem");
+             "org.apache.hadoop.hdfs.ChecksumDistributedFileSystem");
     Random rand = new Random(seed);
     rand.nextBytes(expected);
 

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSOutputSummer.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSOutputSummer.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSOutputSummer.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSOutputSummer.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSOutputSummer.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFSOutputSummer.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFSOutputSummer.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -114,7 +114,7 @@
     conf.setLong("dfs.block.size", BLOCK_SIZE);
     conf.setInt("io.bytes.per.checksum", BYTES_PER_CHECKSUM);
     conf.set("fs.hdfs.impl",
-             "org.apache.hadoop.dfs.ChecksumDistributedFileSystem");      
+             "org.apache.hadoop.hdfs.ChecksumDistributedFileSystem");      
     MiniDFSCluster cluster = new MiniDFSCluster(
         conf, NUM_OF_DATANODES, true, null);
     fileSys = cluster.getFileSystem();

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileAppend.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileAppend.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileAppend.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileAppend.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileAppend.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileAppend.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileAppend.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -30,6 +30,12 @@
 import org.apache.hadoop.fs.FSDataInputStream;
 import org.apache.hadoop.fs.FileUtil.HardLink;
 import org.apache.hadoop.fs.BlockLocation;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.FSDataset;
+import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
 
 /**
  * This class tests the building blocks that are needed to

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCorruption.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCorruption.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCorruption.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCorruption.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCorruption.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCorruption.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCorruption.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.*;
 import junit.framework.*;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreation.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreation.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreation.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreation.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreation.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreation.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreation.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.*;
 import java.net.InetSocketAddress;
@@ -23,6 +23,16 @@
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.*;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.FSConstants;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.LocatedBlocks;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.datanode.FSDataset;
+import org.apache.hadoop.hdfs.server.datanode.SimulatedFSDataset;
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
+import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
 import org.apache.hadoop.io.IOUtils;
 
 import org.apache.commons.logging.impl.Log4JLogger;
@@ -667,7 +677,7 @@
       for(DatanodeInfo datanodeinfo: locatedblock.getLocations()) {
         DataNode datanode = cluster.getDataNode(datanodeinfo.ipcPort);
         FSDataset dataset = (FSDataset)datanode.data;
-        Block b = dataset.getStoredBlock(locatedblock.getBlock().blkid);
+        Block b = dataset.getStoredBlock(locatedblock.getBlock().getBlockId());
         File blockfile = dataset.findBlockFile(b);
         System.out.println("blockfile=" + blockfile);
         if (blockfile != null) {

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationDelete.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationDelete.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationDelete.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationDelete.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationDelete.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationDelete.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationDelete.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 
@@ -24,6 +24,9 @@
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.hdfs.server.namenode.FSNamesystem;
+import org.apache.hadoop.hdfs.server.namenode.LeaseManager;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
 import org.apache.log4j.Level;
 
 public class TestFileCreationDelete extends junit.framework.TestCase {
@@ -91,4 +94,4 @@
       cluster.shutdown();
     }
   }
-}
\ No newline at end of file
+}

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationNamenodeRestart.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationNamenodeRestart.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationNamenodeRestart.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationNamenodeRestart.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationNamenodeRestart.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileCreationNamenodeRestart.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileCreationNamenodeRestart.java Thu Jul  3 15:55:06 2008
@@ -15,10 +15,10 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 public class TestFileCreationNamenodeRestart extends junit.framework.TestCase {
   public void testFileCreationNamenodeRestart() throws Exception {
     new TestFileCreation().xxxtestFileCreationNamenodeRestart();
   }
-}
\ No newline at end of file
+}

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileStatus.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileStatus.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileStatus.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileStatus.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileStatus.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestFileStatus.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestFileStatus.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import junit.framework.TestCase;
 import java.io.*;
@@ -28,6 +28,7 @@
 import org.apache.hadoop.util.StringUtils;
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.BlockLocation;
+import org.apache.hadoop.hdfs.protocol.DFSFileInfo;
 
 /**
  * This class tests the FileStatus API.

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestGetBlocks.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestGetBlocks.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestGetBlocks.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestGetBlocks.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestGetBlocks.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestGetBlocks.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestGetBlocks.java Thu Jul  3 15:55:06 2008
@@ -15,14 +15,19 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.IOException;
 import java.net.InetSocketAddress;
 import java.util.*;
 
 import org.apache.hadoop.conf.Configuration;
-import org.apache.hadoop.dfs.BlocksWithLocations.BlockWithLocations;
+import org.apache.hadoop.hdfs.protocol.Block;
+import org.apache.hadoop.hdfs.protocol.DatanodeInfo;
+import org.apache.hadoop.hdfs.protocol.LocatedBlock;
+import org.apache.hadoop.hdfs.protocol.BlocksWithLocations.BlockWithLocations;
+import org.apache.hadoop.hdfs.server.common.GenerationStamp;
+import org.apache.hadoop.hdfs.server.protocol.NamenodeProtocol;
 import org.apache.hadoop.fs.FSDataOutputStream;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.Path;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSFileSystemContract.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSFileSystemContract.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSFileSystemContract.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSFileSystemContract.java Thu Jul  3 15:55:06 2008
@@ -16,7 +16,7 @@
  * limitations under the License.
  */
 
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystemContractBaseTest;

Copied: hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSServerPorts.java (from r673837, hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSServerPorts.java)
URL: http://svn.apache.org/viewvc/hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSServerPorts.java?p2=hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSServerPorts.java&p1=hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSServerPorts.java&r1=673837&r2=673857&rev=673857&view=diff
==============================================================================
--- hadoop/core/trunk/src/test/org/apache/hadoop/dfs/TestHDFSServerPorts.java (original)
+++ hadoop/core/trunk/src/test/org/apache/hadoop/hdfs/TestHDFSServerPorts.java Thu Jul  3 15:55:06 2008
@@ -15,7 +15,7 @@
  * See the License for the specific language governing permissions and
  * limitations under the License.
  */
-package org.apache.hadoop.dfs;
+package org.apache.hadoop.hdfs;
 
 import java.io.File;
 import java.io.IOException;
@@ -24,6 +24,9 @@
 import org.apache.hadoop.conf.Configuration;
 import org.apache.hadoop.fs.FileSystem;
 import org.apache.hadoop.fs.FileUtil;
+import org.apache.hadoop.hdfs.server.datanode.DataNode;
+import org.apache.hadoop.hdfs.server.namenode.NameNode;
+import org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode;
 import org.apache.hadoop.ipc.RPC;
 
 /**



Mime
View raw message