hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Kitti Nanasi (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-13882) Change dfs.client.block.write.locateFollowingBlock.retries default from 5 to 10
Date Wed, 12 Sep 2018 19:11:00 GMT

    [ https://issues.apache.org/jira/browse/HDFS-13882?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16612618#comment-16612618

Kitti Nanasi commented on HDFS-13882:

Thanks for [~arpitagarwal] and [~xiaochen] for the discussion!

I will not change the default for the retry number then, but I will add a config for the maximum
sleep between retries. I think the default for that maximum sleep could be 60 seconds, if
7 retries usually solves the problem for you with maximum 50 seconds of waiting, 60 seconds
seems like a reasonable maximum to me. What do you think?

> Change dfs.client.block.write.locateFollowingBlock.retries default from 5 to 10
> -------------------------------------------------------------------------------
>                 Key: HDFS-13882
>                 URL: https://issues.apache.org/jira/browse/HDFS-13882
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>    Affects Versions: 3.1.0
>            Reporter: Kitti Nanasi
>            Assignee: Kitti Nanasi
>            Priority: Major
>         Attachments: HDFS-13882.001.patch
> More and more we are seeing cases where customers are running into the java io exception
"Unable to close file because the last block does not have enough number of replicas" on client
file closure. The common workaround is to increase dfs.client.block.write.locateFollowingBlock.retries
from 5 to 10. 

This message was sent by Atlassian JIRA

To unsubscribe, e-mail: hdfs-issues-unsubscribe@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-help@hadoop.apache.org

View raw message