Return-Path: X-Original-To: apmail-hadoop-common-commits-archive@www.apache.org Delivered-To: apmail-hadoop-common-commits-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id B38281742C for ; Sat, 2 May 2015 17:03:28 +0000 (UTC) Received: (qmail 57696 invoked by uid 500); 2 May 2015 17:03:27 -0000 Delivered-To: apmail-hadoop-common-commits-archive@hadoop.apache.org Received: (qmail 57352 invoked by uid 500); 2 May 2015 17:03:27 -0000 Mailing-List: contact common-commits-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-dev@hadoop.apache.org Delivered-To: mailing list common-commits@hadoop.apache.org Received: (qmail 57179 invoked by uid 99); 2 May 2015 17:03:27 -0000 Received: from git1-us-west.apache.org (HELO git1-us-west.apache.org) (140.211.11.23) by apache.org (qpsmtpd/0.29) with ESMTP; Sat, 02 May 2015 17:03:27 +0000 Received: by git1-us-west.apache.org (ASF Mail Server at git1-us-west.apache.org, from userid 33) id 2FB48E17CB; Sat, 2 May 2015 17:03:27 +0000 (UTC) Content-Type: text/plain; charset="us-ascii" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit From: wheat9@apache.org To: common-commits@hadoop.apache.org Date: Sat, 02 May 2015 17:03:29 -0000 Message-Id: <3cf4c189624040dfb77e116cbff92177@git.apache.org> In-Reply-To: <04542ecc87f14b6497edcd3029666800@git.apache.org> References: <04542ecc87f14b6497edcd3029666800@git.apache.org> X-Mailer: ASF-Git Admin Mailer Subject: [3/4] hadoop git commit: HDFS-8249. Separate HdfsConstants into the client and the server side class. Contributed by Haohui Mai. http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java index d26a9a5..09a2d8b 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockPoolSliceStorage.java @@ -24,8 +24,8 @@ import com.google.common.collect.Lists; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.HardLink; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.LayoutVersion; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException; @@ -267,7 +267,7 @@ public class BlockPoolSliceStorage extends Storage { LOG.info("Formatting block pool " + blockpoolID + " directory " + bpSdir.getCurrentDir()); bpSdir.clearDirectory(); // create directory - this.layoutVersion = HdfsConstants.DATANODE_LAYOUT_VERSION; + this.layoutVersion = HdfsServerConstants.DATANODE_LAYOUT_VERSION; this.cTime = nsInfo.getCTime(); this.namespaceID = nsInfo.getNamespaceID(); this.blockpoolID = nsInfo.getBlockPoolID(); @@ -361,7 +361,7 @@ public class BlockPoolSliceStorage extends Storage { } readProperties(sd); checkVersionUpgradable(this.layoutVersion); - assert this.layoutVersion >= HdfsConstants.DATANODE_LAYOUT_VERSION + assert this.layoutVersion >= HdfsServerConstants.DATANODE_LAYOUT_VERSION : "Future version is not allowed"; if (getNamespaceID() != nsInfo.getNamespaceID()) { throw new IOException("Incompatible namespaceIDs in " @@ -375,17 +375,17 @@ public class BlockPoolSliceStorage extends Storage { + nsInfo.getBlockPoolID() + "; datanode blockpoolID = " + blockpoolID); } - if (this.layoutVersion == HdfsConstants.DATANODE_LAYOUT_VERSION + if (this.layoutVersion == HdfsServerConstants.DATANODE_LAYOUT_VERSION && this.cTime == nsInfo.getCTime()) { return; // regular startup } - if (this.layoutVersion > HdfsConstants.DATANODE_LAYOUT_VERSION) { + if (this.layoutVersion > HdfsServerConstants.DATANODE_LAYOUT_VERSION) { int restored = restoreBlockFilesFromTrash(getTrashRootDir(sd)); LOG.info("Restored " + restored + " block files from trash " + "before the layout upgrade. These blocks will be moved to " + "the previous directory during the upgrade"); } - if (this.layoutVersion > HdfsConstants.DATANODE_LAYOUT_VERSION + if (this.layoutVersion > HdfsServerConstants.DATANODE_LAYOUT_VERSION || this.cTime < nsInfo.getCTime()) { doUpgrade(datanode, sd, nsInfo); // upgrade return; @@ -425,7 +425,7 @@ public class BlockPoolSliceStorage extends Storage { } LOG.info("Upgrading block pool storage directory " + bpSd.getRoot() + ".\n old LV = " + this.getLayoutVersion() + "; old CTime = " - + this.getCTime() + ".\n new LV = " + HdfsConstants.DATANODE_LAYOUT_VERSION + + this.getCTime() + ".\n new LV = " + HdfsServerConstants.DATANODE_LAYOUT_VERSION + "; new CTime = " + nsInfo.getCTime()); // get /previous directory String dnRoot = getDataNodeStorageRoot(bpSd.getRoot().getCanonicalPath()); @@ -454,7 +454,7 @@ public class BlockPoolSliceStorage extends Storage { // 3. Create new /current with block files hardlinks and VERSION linkAllBlocks(datanode, bpTmpDir, bpCurDir); - this.layoutVersion = HdfsConstants.DATANODE_LAYOUT_VERSION; + this.layoutVersion = HdfsServerConstants.DATANODE_LAYOUT_VERSION; assert this.namespaceID == nsInfo.getNamespaceID() : "Data-node and name-node layout versions must be the same."; this.cTime = nsInfo.getCTime(); @@ -563,13 +563,13 @@ public class BlockPoolSliceStorage extends Storage { // the namespace state or can be further upgraded to it. // In another word, we can only roll back when ( storedLV >= software LV) // && ( DN.previousCTime <= NN.ctime) - if (!(prevInfo.getLayoutVersion() >= HdfsConstants.DATANODE_LAYOUT_VERSION && + if (!(prevInfo.getLayoutVersion() >= HdfsServerConstants.DATANODE_LAYOUT_VERSION && prevInfo.getCTime() <= nsInfo.getCTime())) { // cannot rollback throw new InconsistentFSStateException(bpSd.getRoot(), "Cannot rollback to a newer state.\nDatanode previous state: LV = " + prevInfo.getLayoutVersion() + " CTime = " + prevInfo.getCTime() + " is newer than the namespace state: LV = " - + HdfsConstants.DATANODE_LAYOUT_VERSION + " CTime = " + nsInfo.getCTime()); + + HdfsServerConstants.DATANODE_LAYOUT_VERSION + " CTime = " + nsInfo.getCTime()); } LOG.info("Rolling back storage directory " + bpSd.getRoot() http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java index c0be956..10692d4 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java @@ -41,13 +41,13 @@ import org.apache.hadoop.fs.FSOutputSummer; import org.apache.hadoop.fs.StorageType; import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.datatransfer.BlockConstructionStage; import org.apache.hadoop.hdfs.protocol.datatransfer.PacketHeader; import org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver; import org.apache.hadoop.hdfs.protocol.datatransfer.PipelineAck; import org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.BlockOpResponseProto; import org.apache.hadoop.hdfs.protocol.proto.DataTransferProtos.Status; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.datanode.fsdataset.ReplicaInputStreams; import org.apache.hadoop.hdfs.server.datanode.fsdataset.ReplicaOutputStreams; import org.apache.hadoop.hdfs.server.protocol.DatanodeRegistration; @@ -246,7 +246,7 @@ class BlockReceiver implements Closeable { out.getClass()); } this.checksumOut = new DataOutputStream(new BufferedOutputStream( - streams.getChecksumOut(), HdfsConstants.SMALL_BUFFER_SIZE)); + streams.getChecksumOut(), HdfsServerConstants.SMALL_BUFFER_SIZE)); // write data chunk header if creating a new replica if (isCreate) { BlockMetadataHeader.writeHeader(checksumOut, diskChecksum); http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java index e76b93a..6b958a2 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java @@ -35,8 +35,8 @@ import java.util.Arrays; import org.apache.commons.logging.Log; import org.apache.hadoop.fs.ChecksumException; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.datatransfer.PacketHeader; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeReference; import org.apache.hadoop.hdfs.server.datanode.fsdataset.LengthInputStream; import org.apache.hadoop.hdfs.util.DataTransferThrottler; @@ -105,7 +105,7 @@ class BlockSender implements java.io.Closeable { */ private static final int MIN_BUFFER_WITH_TRANSFERTO = 64*1024; private static final int TRANSFERTO_BUFFER_SIZE = Math.max( - HdfsConstants.IO_FILE_BUFFER_SIZE, MIN_BUFFER_WITH_TRANSFERTO); + HdfsServerConstants.IO_FILE_BUFFER_SIZE, MIN_BUFFER_WITH_TRANSFERTO); /** the block to read from */ private final ExtendedBlock block; @@ -298,7 +298,7 @@ class BlockSender implements java.io.Closeable { // storage and computes the checksum. if (metaIn.getLength() > BlockMetadataHeader.getHeaderSize()) { checksumIn = new DataInputStream(new BufferedInputStream( - metaIn, HdfsConstants.IO_FILE_BUFFER_SIZE)); + metaIn, HdfsServerConstants.IO_FILE_BUFFER_SIZE)); csum = BlockMetadataHeader.readDataChecksum(checksumIn, block); keepMetaInOpen = true; @@ -747,7 +747,7 @@ class BlockSender implements java.io.Closeable { pktBufSize += checksumSize * maxChunksPerPacket; } else { maxChunksPerPacket = Math.max(1, - numberOfChunks(HdfsConstants.IO_FILE_BUFFER_SIZE)); + numberOfChunks(HdfsServerConstants.IO_FILE_BUFFER_SIZE)); // Packet size includes both checksum and data pktBufSize += (chunkSize + checksumSize) * maxChunksPerPacket; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java index f042dff..22b1520 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNode.java @@ -2156,7 +2156,7 @@ public class DataNode extends ReconfigurableBase unbufIn = saslStreams.in; out = new DataOutputStream(new BufferedOutputStream(unbufOut, - HdfsConstants.SMALL_BUFFER_SIZE)); + HdfsServerConstants.SMALL_BUFFER_SIZE)); in = new DataInputStream(unbufIn); blockSender = new BlockSender(b, 0, b.getNumBytes(), false, false, true, DataNode.this, null, cachingStrategy); http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java index 77fcfed..089e032 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataStorage.java @@ -35,8 +35,8 @@ import org.apache.hadoop.fs.permission.FsPermission; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.HdfsConfiguration; import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.LayoutVersion; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException; @@ -60,7 +60,6 @@ import java.util.Collection; import java.util.Collections; import java.util.Comparator; import java.util.HashMap; -import java.util.HashSet; import java.util.Iterator; import java.util.LinkedList; import java.util.List; @@ -466,7 +465,7 @@ public class DataStorage extends Storage { void recoverTransitionRead(DataNode datanode, NamespaceInfo nsInfo, Collection dataDirs, StartupOption startOpt) throws IOException { if (this.initialized) { - LOG.info("DataNode version: " + HdfsConstants.DATANODE_LAYOUT_VERSION + LOG.info("DataNode version: " + HdfsServerConstants.DATANODE_LAYOUT_VERSION + " and NameNode layout version: " + nsInfo.getLayoutVersion()); this.storageDirs = new ArrayList(dataDirs.size()); // mark DN storage is initialized @@ -509,7 +508,7 @@ public class DataStorage extends Storage { void format(StorageDirectory sd, NamespaceInfo nsInfo, String datanodeUuid) throws IOException { sd.clearDirectory(); // create directory - this.layoutVersion = HdfsConstants.DATANODE_LAYOUT_VERSION; + this.layoutVersion = HdfsServerConstants.DATANODE_LAYOUT_VERSION; this.clusterID = nsInfo.getClusterID(); this.namespaceID = nsInfo.getNamespaceID(); this.cTime = 0; @@ -664,7 +663,7 @@ public class DataStorage extends Storage { } readProperties(sd); checkVersionUpgradable(this.layoutVersion); - assert this.layoutVersion >= HdfsConstants.DATANODE_LAYOUT_VERSION : + assert this.layoutVersion >= HdfsServerConstants.DATANODE_LAYOUT_VERSION : "Future version is not allowed"; boolean federationSupported = @@ -696,13 +695,13 @@ public class DataStorage extends Storage { DatanodeStorage.isValidStorageId(sd.getStorageUuid()); // regular start up. - if (this.layoutVersion == HdfsConstants.DATANODE_LAYOUT_VERSION) { + if (this.layoutVersion == HdfsServerConstants.DATANODE_LAYOUT_VERSION) { createStorageID(sd, !haveValidStorageId); return; // regular startup } // do upgrade - if (this.layoutVersion > HdfsConstants.DATANODE_LAYOUT_VERSION) { + if (this.layoutVersion > HdfsServerConstants.DATANODE_LAYOUT_VERSION) { doUpgrade(datanode, sd, nsInfo); // upgrade createStorageID(sd, !haveValidStorageId); return; @@ -714,7 +713,7 @@ public class DataStorage extends Storage { // failed. throw new IOException("BUG: The stored LV = " + this.getLayoutVersion() + " is newer than the supported LV = " - + HdfsConstants.DATANODE_LAYOUT_VERSION); + + HdfsServerConstants.DATANODE_LAYOUT_VERSION); } /** @@ -749,9 +748,9 @@ public class DataStorage extends Storage { // field and overwrite the file. The upgrade work is handled by // {@link BlockPoolSliceStorage#doUpgrade} LOG.info("Updating layout version from " + layoutVersion + " to " - + HdfsConstants.DATANODE_LAYOUT_VERSION + " for storage " + + HdfsServerConstants.DATANODE_LAYOUT_VERSION + " for storage " + sd.getRoot()); - layoutVersion = HdfsConstants.DATANODE_LAYOUT_VERSION; + layoutVersion = HdfsServerConstants.DATANODE_LAYOUT_VERSION; writeProperties(sd); return; } @@ -759,7 +758,7 @@ public class DataStorage extends Storage { LOG.info("Upgrading storage directory " + sd.getRoot() + ".\n old LV = " + this.getLayoutVersion() + "; old CTime = " + this.getCTime() - + ".\n new LV = " + HdfsConstants.DATANODE_LAYOUT_VERSION + + ".\n new LV = " + HdfsServerConstants.DATANODE_LAYOUT_VERSION + "; new CTime = " + nsInfo.getCTime()); File curDir = sd.getCurrentDir(); @@ -790,7 +789,7 @@ public class DataStorage extends Storage { STORAGE_DIR_CURRENT)); // 4. Write version file under /current - layoutVersion = HdfsConstants.DATANODE_LAYOUT_VERSION; + layoutVersion = HdfsServerConstants.DATANODE_LAYOUT_VERSION; clusterID = nsInfo.getClusterID(); writeProperties(sd); @@ -848,11 +847,11 @@ public class DataStorage extends Storage { // This is a regular startup or a post-federation rollback if (!prevDir.exists()) { if (DataNodeLayoutVersion.supports(LayoutVersion.Feature.FEDERATION, - HdfsConstants.DATANODE_LAYOUT_VERSION)) { - readProperties(sd, HdfsConstants.DATANODE_LAYOUT_VERSION); + HdfsServerConstants.DATANODE_LAYOUT_VERSION)) { + readProperties(sd, HdfsServerConstants.DATANODE_LAYOUT_VERSION); writeProperties(sd); LOG.info("Layout version rolled back to " - + HdfsConstants.DATANODE_LAYOUT_VERSION + " for storage " + + HdfsServerConstants.DATANODE_LAYOUT_VERSION + " for storage " + sd.getRoot()); } return; @@ -862,16 +861,16 @@ public class DataStorage extends Storage { // We allow rollback to a state, which is either consistent with // the namespace state or can be further upgraded to it. - if (!(prevInfo.getLayoutVersion() >= HdfsConstants.DATANODE_LAYOUT_VERSION + if (!(prevInfo.getLayoutVersion() >= HdfsServerConstants.DATANODE_LAYOUT_VERSION && prevInfo.getCTime() <= nsInfo.getCTime())) // cannot rollback throw new InconsistentFSStateException(sd.getRoot(), "Cannot rollback to a newer state.\nDatanode previous state: LV = " + prevInfo.getLayoutVersion() + " CTime = " + prevInfo.getCTime() + " is newer than the namespace state: LV = " - + HdfsConstants.DATANODE_LAYOUT_VERSION + " CTime = " + + HdfsServerConstants.DATANODE_LAYOUT_VERSION + " CTime = " + nsInfo.getCTime()); LOG.info("Rolling back storage directory " + sd.getRoot() - + ".\n target LV = " + HdfsConstants.DATANODE_LAYOUT_VERSION + + ".\n target LV = " + HdfsServerConstants.DATANODE_LAYOUT_VERSION + "; target CTime = " + nsInfo.getCTime()); File tmpDir = sd.getRemovedTmp(); assert !tmpDir.exists() : "removed.tmp directory must not exist."; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java index 3b014e8..de25579 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataXceiver.java @@ -191,7 +191,7 @@ class DataXceiver extends Receiver implements Runnable { socketIn, datanode.getXferAddress().getPort(), datanode.getDatanodeId()); input = new BufferedInputStream(saslStreams.in, - HdfsConstants.SMALL_BUFFER_SIZE); + HdfsServerConstants.SMALL_BUFFER_SIZE); socketOut = saslStreams.out; } catch (InvalidMagicNumberException imne) { if (imne.isHandshake4Encryption()) { @@ -514,7 +514,7 @@ class DataXceiver extends Receiver implements Runnable { long read = 0; OutputStream baseStream = getOutputStream(); DataOutputStream out = new DataOutputStream(new BufferedOutputStream( - baseStream, HdfsConstants.SMALL_BUFFER_SIZE)); + baseStream, HdfsServerConstants.SMALL_BUFFER_SIZE)); checkAccess(out, true, block, blockToken, Op.READ_BLOCK, BlockTokenIdentifier.AccessMode.READ); @@ -658,7 +658,7 @@ class DataXceiver extends Receiver implements Runnable { final DataOutputStream replyOut = new DataOutputStream( new BufferedOutputStream( getOutputStream(), - HdfsConstants.SMALL_BUFFER_SIZE)); + HdfsServerConstants.SMALL_BUFFER_SIZE)); checkAccess(replyOut, isClient, block, blockToken, Op.WRITE_BLOCK, BlockTokenIdentifier.AccessMode.WRITE); @@ -717,7 +717,7 @@ class DataXceiver extends Receiver implements Runnable { unbufMirrorOut = saslStreams.out; unbufMirrorIn = saslStreams.in; mirrorOut = new DataOutputStream(new BufferedOutputStream(unbufMirrorOut, - HdfsConstants.SMALL_BUFFER_SIZE)); + HdfsServerConstants.SMALL_BUFFER_SIZE)); mirrorIn = new DataInputStream(unbufMirrorIn); // Do not propagate allowLazyPersist to downstream DataNodes. @@ -932,7 +932,7 @@ class DataXceiver extends Receiver implements Runnable { .getMetaDataInputStream(block); final DataInputStream checksumIn = new DataInputStream( - new BufferedInputStream(metadataIn, HdfsConstants.IO_FILE_BUFFER_SIZE)); + new BufferedInputStream(metadataIn, HdfsServerConstants.IO_FILE_BUFFER_SIZE)); updateCurrentThreadName("Getting checksum for block " + block); try { //read metadata file @@ -1024,7 +1024,7 @@ class DataXceiver extends Receiver implements Runnable { // set up response stream OutputStream baseStream = getOutputStream(); reply = new DataOutputStream(new BufferedOutputStream( - baseStream, HdfsConstants.SMALL_BUFFER_SIZE)); + baseStream, HdfsServerConstants.SMALL_BUFFER_SIZE)); // send status first writeSuccessWithChecksumInfo(blockSender, reply); @@ -1132,9 +1132,9 @@ class DataXceiver extends Receiver implements Runnable { unbufProxyIn = saslStreams.in; proxyOut = new DataOutputStream(new BufferedOutputStream(unbufProxyOut, - HdfsConstants.SMALL_BUFFER_SIZE)); + HdfsServerConstants.SMALL_BUFFER_SIZE)); proxyReply = new DataInputStream(new BufferedInputStream(unbufProxyIn, - HdfsConstants.IO_FILE_BUFFER_SIZE)); + HdfsServerConstants.IO_FILE_BUFFER_SIZE)); /* send request to the proxy */ IoeDuringCopyBlockOperation = true; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java index f6b058b..62885a9 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DirectoryScanner.java @@ -43,7 +43,7 @@ import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DFSUtil; import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; +import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsDatasetSpi; import org.apache.hadoop.hdfs.server.datanode.fsdataset.FsVolumeSpi; import org.apache.hadoop.util.Daemon; @@ -305,7 +305,7 @@ public class DirectoryScanner implements Runnable { public long getGenStamp() { return metaSuffix != null ? Block.getGenerationStamp( getMetaFile().getName()) : - HdfsConstantsClient.GRANDFATHER_GENERATION_STAMP; + HdfsConstants.GRANDFATHER_GENERATION_STAMP; } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java index 6daf039..414d078 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/BlockPoolSlice.java @@ -41,7 +41,7 @@ import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.BlockListAsLongs; import org.apache.hadoop.hdfs.protocol.BlockListAsLongs.BlockReportReplica; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.datanode.BlockMetadataHeader; import org.apache.hadoop.hdfs.server.datanode.DataStorage; import org.apache.hadoop.hdfs.server.datanode.DatanodeUtil; @@ -612,7 +612,7 @@ class BlockPoolSlice { } checksumIn = new DataInputStream( new BufferedInputStream(new FileInputStream(metaFile), - HdfsConstants.IO_FILE_BUFFER_SIZE)); + HdfsServerConstants.IO_FILE_BUFFER_SIZE)); // read and handle the common header here. For now just a version final DataChecksum checksum = BlockMetadataHeader.readDataChecksum( http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java index b87daec..d72b8bf 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetImpl.java @@ -65,8 +65,8 @@ import org.apache.hadoop.hdfs.protocol.BlockLocalPathInfo; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsBlocksMetadata; import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; import org.apache.hadoop.hdfs.protocol.RecoveryInProgressException; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.ReplicaState; import org.apache.hadoop.hdfs.server.common.Storage; import org.apache.hadoop.hdfs.server.datanode.BlockMetadataHeader; @@ -957,7 +957,7 @@ class FsDatasetImpl implements FsDatasetSpi { } } metaOut = new DataOutputStream(new BufferedOutputStream( - new FileOutputStream(dstMeta), HdfsConstants.SMALL_BUFFER_SIZE)); + new FileOutputStream(dstMeta), HdfsServerConstants.SMALL_BUFFER_SIZE)); BlockMetadataHeader.writeHeader(metaOut, checksum); int offset = 0; @@ -2128,7 +2128,7 @@ class FsDatasetImpl implements FsDatasetSpi { final long diskGS = diskMetaFile != null && diskMetaFile.exists() ? Block.getGenerationStamp(diskMetaFile.getName()) : - HdfsConstantsClient.GRANDFATHER_GENERATION_STAMP; + HdfsConstants.GRANDFATHER_GENERATION_STAMP; if (diskFile == null || !diskFile.exists()) { if (memBlockInfo == null) { @@ -2229,7 +2229,7 @@ class FsDatasetImpl implements FsDatasetSpi { // as the block file, then use the generation stamp from it long gs = diskMetaFile != null && diskMetaFile.exists() && diskMetaFile.getParent().equals(memFile.getParent()) ? diskGS - : HdfsConstantsClient.GRANDFATHER_GENERATION_STAMP; + : HdfsConstants.GRANDFATHER_GENERATION_STAMP; LOG.warn("Updating generation stamp for block " + blockId + " from " + memBlockInfo.getGenerationStamp() + " to " + gs); http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java index 1932706..8c5b4a1 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/fsdataset/impl/FsDatasetUtil.java @@ -24,7 +24,7 @@ import java.util.Arrays; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hdfs.protocol.Block; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; +import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.server.datanode.DatanodeUtil; /** Utility methods. */ @@ -88,7 +88,7 @@ public class FsDatasetUtil { return Block.getGenerationStamp(listdir[j].getName()); } FsDatasetImpl.LOG.warn("Block " + blockFile + " does not have a metafile!"); - return HdfsConstantsClient.GRANDFATHER_GENERATION_STAMP; + return HdfsConstants.GRANDFATHER_GENERATION_STAMP; } /** Find the corresponding meta data file from a given block file */ http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java index f1beb75..b32220c 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/mover/Mover.java @@ -330,7 +330,7 @@ public class Mover { private boolean processFile(String fullPath, HdfsLocatedFileStatus status) { final byte policyId = status.getStoragePolicy(); // currently we ignore files with unspecified storage policy - if (policyId == HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED) { + if (policyId == HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED) { return false; } final BlockStoragePolicy policy = blockStoragePolicies[policyId]; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java index 430f00c..e6560ee 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/BackupNode.java @@ -29,12 +29,12 @@ import org.apache.hadoop.ha.ServiceFailedException; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DFSUtil; import org.apache.hadoop.hdfs.NameNodeProxies; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.UnregisteredNodeException; import org.apache.hadoop.hdfs.protocol.HdfsConstants.SafeModeAction; import org.apache.hadoop.hdfs.protocol.proto.JournalProtocolProtos.JournalProtocolService; import org.apache.hadoop.hdfs.protocolPB.JournalProtocolPB; import org.apache.hadoop.hdfs.protocolPB.JournalProtocolServerSideTranslatorPB; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.Storage; @@ -156,7 +156,7 @@ public class BackupNode extends NameNode { // Backup node should never do lease recovery, // therefore lease hard limit should never expire. namesystem.leaseManager.setLeasePeriod( - HdfsConstants.LEASE_SOFTLIMIT_PERIOD, Long.MAX_VALUE); + HdfsServerConstants.LEASE_SOFTLIMIT_PERIOD, Long.MAX_VALUE); // register with the active name-node registerWith(nsInfo); @@ -414,9 +414,9 @@ public class BackupNode extends NameNode { LOG.error(errorMsg); throw new IOException(errorMsg); } - assert HdfsConstants.NAMENODE_LAYOUT_VERSION == nsInfo.getLayoutVersion() : + assert HdfsServerConstants.NAMENODE_LAYOUT_VERSION == nsInfo.getLayoutVersion() : "Active and backup node layout versions must be the same. Expected: " - + HdfsConstants.NAMENODE_LAYOUT_VERSION + " actual "+ nsInfo.getLayoutVersion(); + + HdfsServerConstants.NAMENODE_LAYOUT_VERSION + " actual "+ nsInfo.getLayoutVersion(); return nsInfo; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogBackupInputStream.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogBackupInputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogBackupInputStream.java index 3649437..689cacc 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogBackupInputStream.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogBackupInputStream.java @@ -21,8 +21,8 @@ import java.io.DataInputStream; import java.io.ByteArrayInputStream; import java.io.IOException; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import com.google.common.base.Preconditions; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; /** * An implementation of the abstract class {@link EditLogInputStream}, @@ -130,12 +130,12 @@ class EditLogBackupInputStream extends EditLogInputStream { @Override public long getFirstTxId() { - return HdfsConstants.INVALID_TXID; + return HdfsServerConstants.INVALID_TXID; } @Override public long getLastTxId() { - return HdfsConstants.INVALID_TXID; + return HdfsServerConstants.INVALID_TXID; } @Override http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogFileInputStream.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogFileInputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogFileInputStream.java index 974860c..3e21c24 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogFileInputStream.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogFileInputStream.java @@ -33,9 +33,9 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.hdfs.DFSConfigKeys; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.LayoutFlags; import org.apache.hadoop.hdfs.protocol.LayoutVersion; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.Storage; import org.apache.hadoop.hdfs.server.namenode.FSEditLogLoader.EditLogValidation; import org.apache.hadoop.hdfs.server.namenode.TransferFsImage.HttpGetFailedException; @@ -85,7 +85,7 @@ public class EditLogFileInputStream extends EditLogInputStream { */ EditLogFileInputStream(File name) throws LogHeaderCorruptException, IOException { - this(name, HdfsConstants.INVALID_TXID, HdfsConstants.INVALID_TXID, false); + this(name, HdfsServerConstants.INVALID_TXID, HdfsServerConstants.INVALID_TXID, false); } /** @@ -203,7 +203,7 @@ public class EditLogFileInputStream extends EditLogInputStream { if ((op != null) && (op.hasTransactionId())) { long txId = op.getTransactionId(); if ((txId >= lastTxId) && - (lastTxId != HdfsConstants.INVALID_TXID)) { + (lastTxId != HdfsServerConstants.INVALID_TXID)) { // // Sometimes, the NameNode crashes while it's writing to the // edit log. In that case, you can end up with an unfinalized edit log @@ -310,7 +310,7 @@ public class EditLogFileInputStream extends EditLogInputStream { // If the header is malformed or the wrong value, this indicates a corruption LOG.warn("Log file " + file + " has no valid header", e); return new FSEditLogLoader.EditLogValidation(0, - HdfsConstants.INVALID_TXID, true); + HdfsServerConstants.INVALID_TXID, true); } try { @@ -331,18 +331,18 @@ public class EditLogFileInputStream extends EditLogInputStream { } catch (LogHeaderCorruptException e) { LOG.warn("Log file " + file + " has no valid header", e); return new FSEditLogLoader.EditLogValidation(0, - HdfsConstants.INVALID_TXID, true); + HdfsServerConstants.INVALID_TXID, true); } long lastPos = 0; - long lastTxId = HdfsConstants.INVALID_TXID; + long lastTxId = HdfsServerConstants.INVALID_TXID; long numValid = 0; try { while (true) { - long txid = HdfsConstants.INVALID_TXID; + long txid = HdfsServerConstants.INVALID_TXID; lastPos = in.getPosition(); try { - if ((txid = in.scanNextOp()) == HdfsConstants.INVALID_TXID) { + if ((txid = in.scanNextOp()) == HdfsServerConstants.INVALID_TXID) { break; } } catch (Throwable t) { @@ -354,7 +354,7 @@ public class EditLogFileInputStream extends EditLogInputStream { FSImage.LOG.warn("After resync, position is " + in.getPosition()); continue; } - if (lastTxId == HdfsConstants.INVALID_TXID || txid > lastTxId) { + if (lastTxId == HdfsServerConstants.INVALID_TXID || txid > lastTxId) { lastTxId = txid; } numValid++; @@ -382,12 +382,12 @@ public class EditLogFileInputStream extends EditLogInputStream { "Reached EOF when reading log header"); } if (verifyLayoutVersion && - (logVersion < HdfsConstants.NAMENODE_LAYOUT_VERSION || // future version + (logVersion < HdfsServerConstants.NAMENODE_LAYOUT_VERSION || // future version logVersion > Storage.LAST_UPGRADABLE_LAYOUT_VERSION)) { // unsupported throw new LogHeaderCorruptException( "Unexpected version of the file system log file: " + logVersion + ". Current version = " - + HdfsConstants.NAMENODE_LAYOUT_VERSION + "."); + + HdfsServerConstants.NAMENODE_LAYOUT_VERSION + "."); } return logVersion; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogInputStream.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogInputStream.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogInputStream.java index ac58616..a4377cd 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogInputStream.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditLogInputStream.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hdfs.server.namenode; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import java.io.Closeable; import java.io.IOException; @@ -112,7 +112,7 @@ public abstract class EditLogInputStream implements Closeable { */ protected long scanNextOp() throws IOException { FSEditLogOp next = readOp(); - return next != null ? next.txid : HdfsConstants.INVALID_TXID; + return next != null ? next.txid : HdfsServerConstants.INVALID_TXID; } /** http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java index f1da3fb..76a9eac 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/EditsDoubleBuffer.java @@ -21,7 +21,7 @@ import java.io.IOException; import java.io.OutputStream; import org.apache.hadoop.classification.InterfaceAudience; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.Writer; import org.apache.hadoop.io.DataOutputBuffer; import org.apache.hadoop.io.IOUtils; @@ -141,7 +141,7 @@ public class EditsDoubleBuffer { } public void writeOp(FSEditLogOp op) throws IOException { - if (firstTxId == HdfsConstants.INVALID_TXID) { + if (firstTxId == HdfsServerConstants.INVALID_TXID) { firstTxId = op.txid; } else { assert op.txid > firstTxId; @@ -153,7 +153,7 @@ public class EditsDoubleBuffer { @Override public DataOutputBuffer reset() { super.reset(); - firstTxId = HdfsConstants.INVALID_TXID; + firstTxId = HdfsServerConstants.INVALID_TXID; numTxns = 0; return this; } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirStatAndListingOp.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirStatAndListingOp.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirStatAndListingOp.java index e77ddae..c636d93 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirStatAndListingOp.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirStatAndListingOp.java @@ -30,7 +30,6 @@ import org.apache.hadoop.hdfs.DFSUtil; import org.apache.hadoop.hdfs.protocol.DirectoryListing; import org.apache.hadoop.hdfs.protocol.FsPermissionExtension; import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; import org.apache.hadoop.hdfs.protocol.HdfsLocatedFileStatus; import org.apache.hadoop.hdfs.protocol.LocatedBlock; @@ -139,7 +138,7 @@ class FSDirStatAndListingOp { } private static byte getStoragePolicyID(byte inodePolicy, byte parentPolicy) { - return inodePolicy != HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED ? inodePolicy : + return inodePolicy != HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED ? inodePolicy : parentPolicy; } @@ -175,7 +174,7 @@ class FSDirStatAndListingOp { if (targetNode == null) return null; byte parentStoragePolicy = isSuperUser ? - targetNode.getStoragePolicyID() : HdfsConstantsClient + targetNode.getStoragePolicyID() : HdfsConstants .BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; if (!targetNode.isDirectory()) { @@ -202,7 +201,7 @@ class FSDirStatAndListingOp { INode cur = contents.get(startChild+i); byte curPolicy = isSuperUser && !cur.isSymlink()? cur.getLocalStoragePolicyID(): - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; INodeAttributes nodeAttrs = getINodeAttributes( fsd, src, cur.getLocalNameBytes(), cur, snapshot); @@ -265,7 +264,7 @@ class FSDirStatAndListingOp { listing[i] = createFileStatus( fsd, sRoot.getLocalNameBytes(), sRoot, nodeAttrs, - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, Snapshot.CURRENT_STATE_ID, false, INodesInPath.fromINode(sRoot)); } @@ -293,8 +292,8 @@ class FSDirStatAndListingOp { } byte policyId = includeStoragePolicy && !i.isSymlink() ? - i.getStoragePolicyID() : HdfsConstantsClient - .BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; + i.getStoragePolicyID() : + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; INodeAttributes nodeAttrs = getINodeAttributes( fsd, path, HdfsFileStatus.EMPTY_NAME, i, src.getPathSnapshotId()); return createFileStatus( @@ -316,7 +315,7 @@ class FSDirStatAndListingOp { if (fsd.getINode4DotSnapshot(srcs) != null) { return new HdfsFileStatus(0, true, 0, 0, 0, 0, null, null, null, null, HdfsFileStatus.EMPTY_NAME, -1L, 0, null, - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED); + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED); } return null; } @@ -330,7 +329,6 @@ class FSDirStatAndListingOp { } } - /** * create an hdfs file status from an inode * http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java index 0189f63..bf538ed 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java @@ -49,7 +49,6 @@ import org.apache.hadoop.hdfs.protocol.EncryptionZone; import org.apache.hadoop.hdfs.protocol.FSLimitException.MaxDirectoryItemsExceededException; import org.apache.hadoop.hdfs.protocol.FSLimitException.PathComponentTooLongException; import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; import org.apache.hadoop.hdfs.protocol.QuotaExceededException; import org.apache.hadoop.hdfs.protocol.SnapshotAccessControlException; @@ -60,6 +59,7 @@ import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguousUnderCon import org.apache.hadoop.hdfs.server.blockmanagement.BlockManager; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo; import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.BlockUCState; import org.apache.hadoop.hdfs.server.namenode.INode.BlocksMapUpdateInfo; import org.apache.hadoop.hdfs.util.ByteArray; @@ -742,7 +742,7 @@ public class FSDirectory implements Closeable { EnumCounters typeSpaceDeltas = new EnumCounters(StorageType.class); // Storage type and its quota are only available when storage policy is set - if (storagePolicyID != HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED) { + if (storagePolicyID != HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED) { BlockStoragePolicy storagePolicy = getBlockManager().getStoragePolicy(storagePolicyID); if (oldRep != newRep) { @@ -884,7 +884,7 @@ public class FSDirectory implements Closeable { /** Verify if the inode name is legal. */ void verifyINodeName(byte[] childName) throws HadoopIllegalArgumentException { - if (Arrays.equals(HdfsConstants.DOT_SNAPSHOT_DIR_BYTES, childName)) { + if (Arrays.equals(HdfsServerConstants.DOT_SNAPSHOT_DIR_BYTES, childName)) { String s = "\"" + HdfsConstants.DOT_SNAPSHOT_DIR + "\" is a reserved name."; if (!namesystem.isImageLoaded()) { s += " Please rename it before upgrade."; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java index df9c585..bda827a 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLog.java @@ -43,9 +43,9 @@ import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo; import org.apache.hadoop.hdfs.protocol.CachePoolInfo; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole; import org.apache.hadoop.hdfs.server.common.Storage.FormatConfirmable; import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory; @@ -159,7 +159,7 @@ public class FSEditLog implements LogsPurgeable { // the first txid of the log that's currently open for writing. // If this value is N, we are currently writing to edits_inprogress_N - private long curSegmentTxId = HdfsConstants.INVALID_TXID; + private long curSegmentTxId = HdfsServerConstants.INVALID_TXID; // the time of printing the statistics to the log file. private long lastPrintTime; @@ -1338,7 +1338,7 @@ public class FSEditLog implements LogsPurgeable { return; } - assert curSegmentTxId == HdfsConstants.INVALID_TXID || // on format this is no-op + assert curSegmentTxId == HdfsServerConstants.INVALID_TXID || // on format this is no-op minTxIdToKeep <= curSegmentTxId : "cannot purge logs older than txid " + minTxIdToKeep + " when current segment starts at " + curSegmentTxId; @@ -1602,7 +1602,7 @@ public class FSEditLog implements LogsPurgeable { EditLogInputStream elis = iter.next(); if (elis.getFirstTxId() > txId) break; long next = elis.getLastTxId(); - if (next == HdfsConstants.INVALID_TXID) { + if (next == HdfsServerConstants.INVALID_TXID) { if (!inProgressOk) { throw new RuntimeException("inProgressOk = false, but " + "selectInputStreams returned an in-progress edit " + http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java index d234cf2..eaa2e77 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogLoader.java @@ -35,16 +35,16 @@ import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.XAttrSetFlag; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; +import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.LocatedBlock; import org.apache.hadoop.hdfs.protocol.Block; import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.HdfsFileStatus; import org.apache.hadoop.hdfs.protocol.LastBlockWithStatus; import org.apache.hadoop.hdfs.protocol.LayoutVersion; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguousUnderConstruction; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.Storage; @@ -302,7 +302,7 @@ public class FSEditLogLoader { long lastInodeId) throws IOException { long inodeId = inodeIdFromOp; - if (inodeId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (inodeId == HdfsConstants.GRANDFATHER_INODE_ID) { if (NameNodeLayoutVersion.supports( LayoutVersion.Feature.ADD_INODE_ID, logVersion)) { throw new IOException("The layout version " + logVersion @@ -322,7 +322,7 @@ public class FSEditLogLoader { @SuppressWarnings("deprecation") private long applyEditLogOp(FSEditLogOp op, FSDirectory fsDir, StartupOption startOpt, int logVersion, long lastInodeId) throws IOException { - long inodeId = HdfsConstantsClient.GRANDFATHER_INODE_ID; + long inodeId = HdfsConstants.GRANDFATHER_INODE_ID; if (LOG.isTraceEnabled()) { LOG.trace("replaying edit log: " + op); } @@ -380,7 +380,7 @@ public class FSEditLogLoader { if (toAddRetryCache) { HdfsFileStatus stat = FSDirStatAndListingOp.createFileStatusForEditLog( fsNamesys.dir, path, HdfsFileStatus.EMPTY_NAME, newFile, - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, Snapshot.CURRENT_STATE_ID, + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, Snapshot.CURRENT_STATE_ID, false, iip); fsNamesys.addCacheEntryWithPayload(addCloseOp.rpcClientId, addCloseOp.rpcCallId, stat); @@ -400,7 +400,7 @@ public class FSEditLogLoader { HdfsFileStatus stat = FSDirStatAndListingOp.createFileStatusForEditLog( fsNamesys.dir, path, HdfsFileStatus.EMPTY_NAME, newFile, - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, Snapshot.CURRENT_STATE_ID, false, iip); fsNamesys.addCacheEntryWithPayload(addCloseOp.rpcClientId, addCloseOp.rpcCallId, new LastBlockWithStatus(lb, stat)); @@ -473,7 +473,7 @@ public class FSEditLogLoader { if (toAddRetryCache) { HdfsFileStatus stat = FSDirStatAndListingOp.createFileStatusForEditLog( fsNamesys.dir, path, HdfsFileStatus.EMPTY_NAME, file, - HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, + HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED, Snapshot.CURRENT_STATE_ID, false, iip); fsNamesys.addCacheEntryWithPayload(appendOp.rpcClientId, appendOp.rpcCallId, new LastBlockWithStatus(lb, stat)); @@ -1095,7 +1095,7 @@ public class FSEditLogLoader { // The editlog must be emptied by restarting the namenode, before proceeding // with the upgrade. if (Storage.is203LayoutVersion(logVersion) - && logVersion != HdfsConstants.NAMENODE_LAYOUT_VERSION) { + && logVersion != HdfsServerConstants.NAMENODE_LAYOUT_VERSION) { String msg = "During upgrade failed to load the editlog version " + logVersion + " from release 0.20.203. Please go back to the old " + " release and restart the namenode. This empties the editlog " @@ -1112,7 +1112,7 @@ public class FSEditLogLoader { */ static EditLogValidation validateEditLog(EditLogInputStream in) { long lastPos = 0; - long lastTxId = HdfsConstants.INVALID_TXID; + long lastTxId = HdfsServerConstants.INVALID_TXID; long numValid = 0; FSEditLogOp op = null; while (true) { @@ -1129,7 +1129,7 @@ public class FSEditLogLoader { FSImage.LOG.warn("After resync, position is " + in.getPosition()); continue; } - if (lastTxId == HdfsConstants.INVALID_TXID + if (lastTxId == HdfsServerConstants.INVALID_TXID || op.getTransactionId() > lastTxId) { lastTxId = op.getTransactionId(); } @@ -1140,7 +1140,7 @@ public class FSEditLogLoader { static EditLogValidation scanEditLog(EditLogInputStream in) { long lastPos = 0; - long lastTxId = HdfsConstants.INVALID_TXID; + long lastTxId = HdfsServerConstants.INVALID_TXID; long numValid = 0; FSEditLogOp op = null; while (true) { @@ -1157,7 +1157,7 @@ public class FSEditLogLoader { FSImage.LOG.warn("After resync, position is " + in.getPosition()); continue; } - if (lastTxId == HdfsConstants.INVALID_TXID + if (lastTxId == HdfsServerConstants.INVALID_TXID || op.getTransactionId() > lastTxId) { lastTxId = op.getTransactionId(); } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java index 0daf12f..c8e565e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSEditLogOp.java @@ -101,13 +101,13 @@ import org.apache.hadoop.hdfs.protocol.CacheDirectiveInfo; import org.apache.hadoop.hdfs.protocol.CachePoolInfo; import org.apache.hadoop.hdfs.protocol.ClientProtocol; import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; import org.apache.hadoop.hdfs.protocol.LayoutVersion; import org.apache.hadoop.hdfs.protocol.LayoutVersion.Feature; import org.apache.hadoop.hdfs.protocol.proto.AclProtos.AclEditLogProto; import org.apache.hadoop.hdfs.protocol.proto.XAttrProtos.XAttrEditLogProto; import org.apache.hadoop.hdfs.protocolPB.PBHelper; import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.util.XMLUtils; import org.apache.hadoop.hdfs.util.XMLUtils.InvalidXmlException; import org.apache.hadoop.hdfs.util.XMLUtils.Stanza; @@ -148,7 +148,7 @@ public abstract class FSEditLogOp { int rpcCallId; final void reset() { - txid = HdfsConstants.INVALID_TXID; + txid = HdfsServerConstants.INVALID_TXID; rpcClientId = RpcConstants.DUMMY_CLIENT_ID; rpcCallId = RpcConstants.INVALID_CALL_ID; resetSubFields(); @@ -241,16 +241,16 @@ public abstract class FSEditLogOp { } public long getTransactionId() { - Preconditions.checkState(txid != HdfsConstants.INVALID_TXID); + Preconditions.checkState(txid != HdfsServerConstants.INVALID_TXID); return txid; } public String getTransactionIdStr() { - return (txid == HdfsConstants.INVALID_TXID) ? "(none)" : "" + txid; + return (txid == HdfsServerConstants.INVALID_TXID) ? "(none)" : "" + txid; } public boolean hasTransactionId() { - return (txid != HdfsConstants.INVALID_TXID); + return (txid != HdfsServerConstants.INVALID_TXID); } public void setTransactionId(long txid) { @@ -433,7 +433,7 @@ public abstract class FSEditLogOp { private AddCloseOp(FSEditLogOpCodes opCode) { super(opCode); - storagePolicyId = HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; + storagePolicyId = HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; assert(opCode == OP_ADD || opCode == OP_CLOSE || opCode == OP_APPEND); } @@ -577,7 +577,7 @@ public abstract class FSEditLogOp { this.inodeId = in.readLong(); } else { // The inodeId should be updated when this editLogOp is applied - this.inodeId = HdfsConstantsClient.GRANDFATHER_INODE_ID; + this.inodeId = HdfsConstants.GRANDFATHER_INODE_ID; } if ((-17 < logVersion && length != 4) || (logVersion <= -17 && length != 5 && !NameNodeLayoutVersion.supports( @@ -635,7 +635,7 @@ public abstract class FSEditLogOp { NameNodeLayoutVersion.Feature.BLOCK_STORAGE_POLICY, logVersion)) { this.storagePolicyId = FSImageSerialization.readByte(in); } else { - this.storagePolicyId = HdfsConstantsClient.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; + this.storagePolicyId = HdfsConstants.BLOCK_STORAGE_POLICY_ID_UNSPECIFIED; } // read clientId and callId readRpcIds(in, logVersion); @@ -1652,7 +1652,7 @@ public abstract class FSEditLogOp { this.inodeId = FSImageSerialization.readLong(in); } else { // This id should be updated when this editLogOp is applied - this.inodeId = HdfsConstantsClient.GRANDFATHER_INODE_ID; + this.inodeId = HdfsConstants.GRANDFATHER_INODE_ID; } this.path = FSImageSerialization.readString(in); if (NameNodeLayoutVersion.supports( @@ -2545,7 +2545,7 @@ public abstract class FSEditLogOp { this.inodeId = FSImageSerialization.readLong(in); } else { // This id should be updated when the editLogOp is applied - this.inodeId = HdfsConstantsClient.GRANDFATHER_INODE_ID; + this.inodeId = HdfsConstants.GRANDFATHER_INODE_ID; } this.path = FSImageSerialization.readString(in); this.value = FSImageSerialization.readString(in); @@ -4689,7 +4689,7 @@ public abstract class FSEditLogOp { // Read the txid op.setTransactionId(in.readLong()); } else { - op.setTransactionId(HdfsConstants.INVALID_TXID); + op.setTransactionId(HdfsServerConstants.INVALID_TXID); } op.readFields(in, logVersion); @@ -4712,13 +4712,13 @@ public abstract class FSEditLogOp { try { opCodeByte = in.readByte(); // op code } catch (EOFException e) { - return HdfsConstants.INVALID_TXID; + return HdfsServerConstants.INVALID_TXID; } FSEditLogOpCodes opCode = FSEditLogOpCodes.fromByte(opCodeByte); if (opCode == OP_INVALID) { verifyTerminator(); - return HdfsConstants.INVALID_TXID; + return HdfsServerConstants.INVALID_TXID; } int length = in.readInt(); // read the length of the op @@ -4730,7 +4730,7 @@ public abstract class FSEditLogOp { return txid; } else { FSEditLogOp op = decodeOp(); - return op == null ? HdfsConstants.INVALID_TXID : op.getTransactionId(); + return op == null ? HdfsServerConstants.INVALID_TXID : op.getTransactionId(); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java index 433b05a..6a54967 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImage.java @@ -44,9 +44,9 @@ import org.apache.hadoop.fs.StorageType; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.hdfs.DFSUtil; import org.apache.hadoop.hdfs.HAUtil; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.LayoutVersion; import org.apache.hadoop.hdfs.server.blockmanagement.BlockStoragePolicySuite; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; @@ -230,7 +230,7 @@ public class FSImage implements Closeable { if (startOpt == StartupOption.METADATAVERSION) { System.out.println("HDFS Image Version: " + layoutVersion); System.out.println("Software format version: " + - HdfsConstants.NAMENODE_LAYOUT_VERSION); + HdfsServerConstants.NAMENODE_LAYOUT_VERSION); return false; } @@ -241,11 +241,11 @@ public class FSImage implements Closeable { && startOpt != StartupOption.UPGRADEONLY && !RollingUpgradeStartupOption.STARTED.matches(startOpt) && layoutVersion < Storage.LAST_PRE_UPGRADE_LAYOUT_VERSION - && layoutVersion != HdfsConstants.NAMENODE_LAYOUT_VERSION) { + && layoutVersion != HdfsServerConstants.NAMENODE_LAYOUT_VERSION) { throw new IOException( "\nFile system image contains an old layout version " + storage.getLayoutVersion() + ".\nAn upgrade to version " - + HdfsConstants.NAMENODE_LAYOUT_VERSION + " is required.\n" + + HdfsServerConstants.NAMENODE_LAYOUT_VERSION + " is required.\n" + "Please restart NameNode with the \"" + RollingUpgradeStartupOption.STARTED.getOptionString() + "\" option if a rolling upgrade is already started;" @@ -397,7 +397,7 @@ public class FSImage implements Closeable { long oldCTime = storage.getCTime(); storage.cTime = now(); // generate new cTime for the state int oldLV = storage.getLayoutVersion(); - storage.layoutVersion = HdfsConstants.NAMENODE_LAYOUT_VERSION; + storage.layoutVersion = HdfsServerConstants.NAMENODE_LAYOUT_VERSION; List errorSDs = Collections.synchronizedList(new ArrayList()); @@ -458,11 +458,11 @@ public class FSImage implements Closeable { boolean canRollback = false; FSImage prevState = new FSImage(conf); try { - prevState.getStorage().layoutVersion = HdfsConstants.NAMENODE_LAYOUT_VERSION; + prevState.getStorage().layoutVersion = HdfsServerConstants.NAMENODE_LAYOUT_VERSION; for (Iterator it = storage.dirIterator(false); it.hasNext();) { StorageDirectory sd = it.next(); if (!NNUpgradeUtil.canRollBack(sd, storage, prevState.getStorage(), - HdfsConstants.NAMENODE_LAYOUT_VERSION)) { + HdfsServerConstants.NAMENODE_LAYOUT_VERSION)) { continue; } LOG.info("Can perform rollback for " + sd); @@ -473,7 +473,7 @@ public class FSImage implements Closeable { // If HA is enabled, check if the shared log can be rolled back as well. editLog.initJournalsForWrite(); boolean canRollBackSharedEditLog = editLog.canRollBackSharedLog( - prevState.getStorage(), HdfsConstants.NAMENODE_LAYOUT_VERSION); + prevState.getStorage(), HdfsServerConstants.NAMENODE_LAYOUT_VERSION); if (canRollBackSharedEditLog) { LOG.info("Can perform rollback for shared edit log."); canRollback = true; @@ -839,7 +839,7 @@ public class FSImage implements Closeable { lastAppliedTxId = loader.getLastAppliedTxId(); } // If we are in recovery mode, we may have skipped over some txids. - if (editIn.getLastTxId() != HdfsConstants.INVALID_TXID) { + if (editIn.getLastTxId() != HdfsServerConstants.INVALID_TXID) { lastAppliedTxId = editIn.getLastTxId(); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java index 906213e..d1d007f 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormat.java @@ -55,6 +55,7 @@ import org.apache.hadoop.hdfs.protocol.LayoutVersion.Feature; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguous; import org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoContiguousUnderConstruction; import org.apache.hadoop.hdfs.server.blockmanagement.BlockManager; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.StartupOption; import org.apache.hadoop.hdfs.server.common.InconsistentFSStateException; import org.apache.hadoop.hdfs.server.namenode.snapshot.DirectoryWithSnapshotFeature; @@ -1046,10 +1047,10 @@ public class FSImageFormat { @VisibleForTesting public static void useDefaultRenameReservedPairs() { renameReservedMap.clear(); - for (String key: HdfsConstants.RESERVED_PATH_COMPONENTS) { + for (String key: HdfsServerConstants.RESERVED_PATH_COMPONENTS) { renameReservedMap.put( key, - key + "." + HdfsConstants.NAMENODE_LAYOUT_VERSION + "." + key + "." + HdfsServerConstants.NAMENODE_LAYOUT_VERSION + "." + "UPGRADE_RENAMED"); } } @@ -1147,7 +1148,7 @@ public class FSImageFormat { final int layoutVersion) { // If the LV doesn't support snapshots, we're doing an upgrade if (!NameNodeLayoutVersion.supports(Feature.SNAPSHOT, layoutVersion)) { - if (Arrays.equals(component, HdfsConstants.DOT_SNAPSHOT_DIR_BYTES)) { + if (Arrays.equals(component, HdfsServerConstants.DOT_SNAPSHOT_DIR_BYTES)) { Preconditions.checkArgument( renameReservedMap.containsKey(HdfsConstants.DOT_SNAPSHOT_DIR), RESERVED_ERROR_MSG); http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatProtobuf.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatProtobuf.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatProtobuf.java index 3ee848a..24afcae 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatProtobuf.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatProtobuf.java @@ -42,11 +42,11 @@ import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.CacheDirectiveInfoProto; import org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos.CachePoolInfoProto; import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSecretManager; import org.apache.hadoop.hdfs.server.blockmanagement.BlockIdManager; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.CacheManagerSection; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.FileSummary; import org.apache.hadoop.hdfs.server.namenode.FsImageProto.NameSystemSection; @@ -143,7 +143,7 @@ public final class FSImageFormatProtobuf { private long imgTxId; /** * Whether the image's layout version must be the same with - * {@link HdfsConstants#NAMENODE_LAYOUT_VERSION}. This is only set to true + * {@link HdfsServerConstants#NAMENODE_LAYOUT_VERSION}. This is only set to true * when we're doing (rollingUpgrade rollback). */ private final boolean requireSameLayoutVersion; @@ -192,10 +192,10 @@ public final class FSImageFormatProtobuf { } FileSummary summary = FSImageUtil.loadSummary(raFile); if (requireSameLayoutVersion && summary.getLayoutVersion() != - HdfsConstants.NAMENODE_LAYOUT_VERSION) { + HdfsServerConstants.NAMENODE_LAYOUT_VERSION) { throw new IOException("Image version " + summary.getLayoutVersion() + " is not equal to the software version " + - HdfsConstants.NAMENODE_LAYOUT_VERSION); + HdfsServerConstants.NAMENODE_LAYOUT_VERSION); } FileChannel channel = fin.getChannel(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java index d5243e0..3cd5ae1 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImagePreTransactionalStorageInspector.java @@ -34,7 +34,7 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; import org.apache.hadoop.fs.FileUtil; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory; import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeDirType; import org.apache.hadoop.hdfs.server.namenode.NNStorage.NameNodeFile; @@ -180,7 +180,7 @@ class FSImagePreTransactionalStorageInspector extends FSImageStorageInspector { FSImageFile file = new FSImageFile(latestNameSD, NNStorage.getStorageFile(latestNameSD, NameNodeFile.IMAGE), - HdfsConstants.INVALID_TXID); + HdfsServerConstants.INVALID_TXID); LinkedList ret = new LinkedList(); ret.add(file); return ret; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageStorageInspector.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageStorageInspector.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageStorageInspector.java index cb1351b..dfda34e 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageStorageInspector.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageStorageInspector.java @@ -21,9 +21,9 @@ import java.io.File; import java.io.IOException; import java.util.List; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.classification.InterfaceStability; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory; /** @@ -69,7 +69,7 @@ abstract class FSImageStorageInspector { private final File file; FSImageFile(StorageDirectory sd, File file, long txId) { - assert txId >= 0 || txId == HdfsConstants.INVALID_TXID + assert txId >= 0 || txId == HdfsServerConstants.INVALID_TXID : "Invalid txid on " + file +": " + txId; this.sd = sd; http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java index 809d594..72e1d32 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSNamesystem.java @@ -184,7 +184,6 @@ import org.apache.hadoop.hdfs.protocol.DirectoryListing; import org.apache.hadoop.hdfs.protocol.EncryptionZone; import org.apache.hadoop.hdfs.protocol.ExtendedBlock; import org.apache.hadoop.hdfs.protocol.HdfsConstants; -import org.apache.hadoop.hdfs.protocol.HdfsConstantsClient; import org.apache.hadoop.hdfs.protocol.LastBlockWithStatus; import org.apache.hadoop.hdfs.protocol.HdfsConstants.DatanodeReportType; import org.apache.hadoop.hdfs.protocol.HdfsConstants.SafeModeAction; @@ -212,6 +211,7 @@ import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeDescriptor; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStatistics; import org.apache.hadoop.hdfs.server.blockmanagement.DatanodeStorageInfo; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.BlockUCState; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.RollingUpgradeStartupOption; @@ -1021,7 +1021,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, private void updateStorageVersionForRollingUpgrade(final long layoutVersion, StartupOption startOpt) throws IOException { boolean rollingStarted = RollingUpgradeStartupOption.STARTED - .matches(startOpt) && layoutVersion > HdfsConstants + .matches(startOpt) && layoutVersion > HdfsServerConstants .NAMENODE_LAYOUT_VERSION; boolean rollingRollback = RollingUpgradeStartupOption.ROLLBACK .matches(startOpt); @@ -3234,7 +3234,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, Block previousBlock = ExtendedBlock.getLocalBlock(previous); final INode inode; final INodesInPath iip; - if (fileId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (fileId == HdfsConstants.GRANDFATHER_INODE_ID) { // Older clients may not have given us an inode ID to work with. // In this case, we have to try to resolve the path and hope it // hasn't changed or been deleted since the file was opened for write. @@ -3354,7 +3354,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, //check lease final INode inode; - if (fileId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (fileId == HdfsConstants.GRANDFATHER_INODE_ID) { // Older clients may not have given us an inode ID to work with. // In this case, we have to try to resolve the path and hope it // hasn't changed or been deleted since the file was opened for write. @@ -3411,7 +3411,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, final INode inode; final INodesInPath iip; - if (fileId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (fileId == HdfsConstants.GRANDFATHER_INODE_ID) { // Older clients may not have given us an inode ID to work with. // In this case, we have to try to resolve the path and hope it // hasn't changed or been deleted since the file was opened for write. @@ -3530,7 +3530,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, final INodesInPath iip; INode inode = null; try { - if (fileId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (fileId == HdfsConstants.GRANDFATHER_INODE_ID) { // Older clients may not have given us an inode ID to work with. // In this case, we have to try to resolve the path and hope it // hasn't changed or been deleted since the file was opened for write. @@ -3981,7 +3981,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, checkNameNodeSafeMode("Cannot fsync file " + src); src = dir.resolvePath(pc, src, pathComponents); final INode inode; - if (fileId == HdfsConstantsClient.GRANDFATHER_INODE_ID) { + if (fileId == HdfsConstants.GRANDFATHER_INODE_ID) { // Older clients may not have given us an inode ID to work with. // In this case, we have to try to resolve the path and hope it // hasn't changed or been deleted since the file was opened for write. @@ -4439,7 +4439,8 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, //remove lease, close file finalizeINodeFileUnderConstruction(src, pendingFile, - Snapshot.findLatestSnapshot(pendingFile, Snapshot.CURRENT_STATE_ID)); + Snapshot.findLatestSnapshot(pendingFile, + Snapshot.CURRENT_STATE_ID)); return src; } @@ -5731,7 +5732,7 @@ public class FSNamesystem implements Namesystem, FSNamesystemMBean, SafeModeInfo safeMode = this.safeMode; if (safeMode == null) return; - safeMode.setBlockTotal((int)getCompleteBlocksTotal()); + safeMode.setBlockTotal((int) getCompleteBlocksTotal()); } /** http://git-wip-us.apache.org/repos/asf/hadoop/blob/6ae2a0d0/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java ---------------------------------------------------------------------- diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java index b760dd8..ebd7475 100644 --- a/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java +++ b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FileJournalManager.java @@ -32,7 +32,7 @@ import java.util.regex.Pattern; import org.apache.hadoop.classification.InterfaceAudience; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.FileUtil; -import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.hadoop.hdfs.server.common.HdfsServerConstants; import org.apache.hadoop.hdfs.server.common.Storage; import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NodeType; import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory; @@ -286,7 +286,7 @@ public class FileJournalManager implements JournalManager { try { long startTxId = Long.parseLong(inProgressEditsMatch.group(1)); ret.add( - new EditLogFile(f, startTxId, HdfsConstants.INVALID_TXID, true)); + new EditLogFile(f, startTxId, HdfsServerConstants.INVALID_TXID, true)); continue; } catch (NumberFormatException nfe) { LOG.error("In-progress edits file " + f + " has improperly " + @@ -301,7 +301,7 @@ public class FileJournalManager implements JournalManager { if (staleInprogressEditsMatch.matches()) { try { long startTxId = Long.parseLong(staleInprogressEditsMatch.group(1)); - ret.add(new EditLogFile(f, startTxId, HdfsConstants.INVALID_TXID, + ret.add(new EditLogFile(f, startTxId, HdfsServerConstants.INVALID_TXID, true)); continue; } catch (NumberFormatException nfe) { @@ -348,7 +348,7 @@ public class FileJournalManager implements JournalManager { } } if (elf.lastTxId < fromTxId) { - assert elf.lastTxId != HdfsConstants.INVALID_TXID; + assert elf.lastTxId != HdfsServerConstants.INVALID_TXID; if (LOG.isDebugEnabled()) { LOG.debug("passing over " + elf + " because it ends at " + elf.lastTxId + ", but we only care about transactions " + @@ -391,7 +391,7 @@ public class FileJournalManager implements JournalManager { throw new CorruptionException("In-progress edit log file is corrupt: " + elf); } - if (elf.getLastTxId() == HdfsConstants.INVALID_TXID) { + if (elf.getLastTxId() == HdfsServerConstants.INVALID_TXID) { // If the file has a valid header (isn't corrupt) but contains no // transactions, we likely just crashed after opening the file and // writing the header, but before syncing any transactions. Safe to @@ -480,19 +480,19 @@ public class FileJournalManager implements JournalManager { EditLogFile(File file, long firstTxId, long lastTxId) { this(file, firstTxId, lastTxId, false); - assert (lastTxId != HdfsConstants.INVALID_TXID) + assert (lastTxId != HdfsServerConstants.INVALID_TXID) && (lastTxId >= firstTxId); } EditLogFile(File file, long firstTxId, long lastTxId, boolean isInProgress) { - assert (lastTxId == HdfsConstants.INVALID_TXID && isInProgress) - || (lastTxId != HdfsConstants.INVALID_TXID && lastTxId >= firstTxId); - assert (firstTxId > 0) || (firstTxId == HdfsConstants.INVALID_TXID); + assert (lastTxId == HdfsServerConstants.INVALID_TXID && isInProgress) + || (lastTxId != HdfsServerConstants.INVALID_TXID && lastTxId >= firstTxId); + assert (firstTxId > 0) || (firstTxId == HdfsServerConstants.INVALID_TXID); assert file != null; Preconditions.checkArgument(!isInProgress || - lastTxId == HdfsConstants.INVALID_TXID); + lastTxId == HdfsServerConstants.INVALID_TXID); this.firstTxId = firstTxId; this.lastTxId = lastTxId; @@ -552,7 +552,7 @@ public class FileJournalManager implements JournalManager { } public void moveAsideEmptyFile() throws IOException { - assert lastTxId == HdfsConstants.INVALID_TXID; + assert lastTxId == HdfsServerConstants.INVALID_TXID; renameSelf(".empty"); }