hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Eli Collins (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HDFS-611) Heartbeats times from Datanodes increase when there are plenty of blocks to delete
Date Thu, 10 Sep 2009 22:39:57 GMT

    [ https://issues.apache.org/jira/browse/HDFS-611?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12753869#action_12753869
] 

Eli Collins commented on HDFS-611:
----------------------------------

Why is the heartbeat thread involved with block deletion?  

Per the fifth comment in HDFS-599 can someone explain in general why the failure detection
code is not encapsulated from the rest of the daemon? If it were a separate module it seems
that we wouldn't have to worry about RPC priorities, JVM GC, block deletion, etc interfering
with it.

Thanks,
Eli

> Heartbeats times from Datanodes increase when there are plenty of blocks to delete
> ----------------------------------------------------------------------------------
>
>                 Key: HDFS-611
>                 URL: https://issues.apache.org/jira/browse/HDFS-611
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: data-node
>            Reporter: dhruba borthakur
>            Assignee: dhruba borthakur
>
> I am seeing that when we delete a large directory that has plenty of blocks, the heartbeat
times from datanodes increase significantly from the normal value of 3 seconds to as large
as 50 seconds or so. The heartbeat thread in the Datanode deletes a bunch of blocks sequentially,
this causes the heartbeat times to increase.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message