hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Kihwal Lee (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-7645) Rolling upgrade is restoring blocks from trash multiple times
Date Thu, 19 Mar 2015 15:36:40 GMT

    [ https://issues.apache.org/jira/browse/HDFS-7645?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14369558#comment-14369558
] 

Kihwal Lee commented on HDFS-7645:
----------------------------------

How will it work when old datanodes are talking to a new namenode? When a RU is finalized
in the middle, the old datanodes will still think the RU is in progress. If we assume they
will get upgraded soon after, this might not be a problem. Otherwise their trash will grow.
 In the opposite situation, newer datanodes will not clear trash until the namenode is upgraded
and the upgrade finalized. But this should not happen if the normal upgrade procedure is followed.
May be it might happen in some cases for federated clusters.  In all cases, the delay in clearing
trash will be temporary.  Agree?

The rest of patch looks fine.

> Rolling upgrade is restoring blocks from trash multiple times
> -------------------------------------------------------------
>
>                 Key: HDFS-7645
>                 URL: https://issues.apache.org/jira/browse/HDFS-7645
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: datanode
>    Affects Versions: 2.6.0
>            Reporter: Nathan Roberts
>            Assignee: Keisuke Ogiwara
>         Attachments: HDFS-7645.01.patch, HDFS-7645.02.patch, HDFS-7645.03.patch, HDFS-7645.04.patch,
HDFS-7645.05.patch
>
>
> When performing an HDFS rolling upgrade, the trash directory is getting restored twice
when under normal circumstances it shouldn't need to be restored at all. iiuc, the only time
these blocks should be restored is if we need to rollback a rolling upgrade. 
> On a busy cluster, this can cause significant and unnecessary block churn both on the
datanodes, and more importantly in the namenode.
> The two times this happens are:
> 1) restart of DN onto new software
> {code}
>   private void doTransition(DataNode datanode, StorageDirectory sd,
>       NamespaceInfo nsInfo, StartupOption startOpt) throws IOException {
>     if (startOpt == StartupOption.ROLLBACK && sd.getPreviousDir().exists()) {
>       Preconditions.checkState(!getTrashRootDir(sd).exists(),
>           sd.getPreviousDir() + " and " + getTrashRootDir(sd) + " should not " +
>           " both be present.");
>       doRollback(sd, nsInfo); // rollback if applicable
>     } else {
>       // Restore all the files in the trash. The restored files are retained
>       // during rolling upgrade rollback. They are deleted during rolling
>       // upgrade downgrade.
>       int restored = restoreBlockFilesFromTrash(getTrashRootDir(sd));
>       LOG.info("Restored " + restored + " block files from trash.");
>     }
> {code}
> 2) When heartbeat response no longer indicates a rollingupgrade is in progress
> {code}
>   /**
>    * Signal the current rolling upgrade status as indicated by the NN.
>    * @param inProgress true if a rolling upgrade is in progress
>    */
>   void signalRollingUpgrade(boolean inProgress) throws IOException {
>     String bpid = getBlockPoolId();
>     if (inProgress) {
>       dn.getFSDataset().enableTrash(bpid);
>       dn.getFSDataset().setRollingUpgradeMarker(bpid);
>     } else {
>       dn.getFSDataset().restoreTrash(bpid);
>       dn.getFSDataset().clearRollingUpgradeMarker(bpid);
>     }
>   }
> {code}
> HDFS-6800 and HDFS-6981 were modifying this behavior making it not completely clear whether
this is somehow intentional. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Mime
View raw message