hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Ming Ma (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-9314) Improve BlockPlacementPolicyDefault's picking of excess replicas
Date Sat, 21 Nov 2015 02:39:11 GMT

    [ https://issues.apache.org/jira/browse/HDFS-9314?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15020186#comment-15020186

Ming Ma commented on HDFS-9314:


* If there are two replicas node1(rack 1) and node2(rack 2), {{pickupReplicaSet}} should still
return both nodes. Maybe it is useful to have a new test for this special case.
* Nit: Comment {{storages ({storages[3]}, {storages[7], storages[8]}) are on different racks
(r1, r2, r3)}} should be {{storages ({storages[3]}, {storages[7], storages[8]}) are on different
racks (r2, r3, r3)}}.

> Improve BlockPlacementPolicyDefault's picking of excess replicas
> ----------------------------------------------------------------
>                 Key: HDFS-9314
>                 URL: https://issues.apache.org/jira/browse/HDFS-9314
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: Ming Ma
>            Assignee: Xiao Chen
>         Attachments: HDFS-9314.001.patch, HDFS-9314.002.patch, HDFS-9314.003.patch, HDFS-9314.004.patch
> The test case used in HDFS-9313 identified NullPointerException as well as the limitation
of excess replica picking. If the current replicas are on {SSD(rack r1), DISK(rack 2), DISK(rack
3), DISK(rack 3)} and the storage policy changes to HOT_STORAGE_POLICY_ID, BlockPlacementPolicyDefault's
won't be able to delete SSD replica.

This message was sent by Atlassian JIRA

View raw message