hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "stack (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HDFS-927) DFSInputStream retries too many times for new block locations
Date Mon, 01 Feb 2010 23:10:18 GMT

    [ https://issues.apache.org/jira/browse/HDFS-927?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12828324#action_12828324

stack commented on HDFS-927:

.bq I actually mean "per read per block", i.e. within a singe read, there are 3 retries for
each block.

That would be better in that with the current patch, on a 3 block file, if we hiccupped the
first read on each block, we'd trip the failures count though if we'd been counting on a block
basis, the read would have gone through.

That said, I'd be fine with Todds patch -- its nice and clean, semantically and code-wise
-- going in and in a new issue working on the Tsz suggested improvement.

> DFSInputStream retries too many times for new block locations
> -------------------------------------------------------------
>                 Key: HDFS-927
>                 URL: https://issues.apache.org/jira/browse/HDFS-927
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs client
>    Affects Versions: 0.21.0, 0.22.0
>            Reporter: Todd Lipcon
>            Assignee: Todd Lipcon
>            Priority: Critical
>         Attachments: hdfs-927.txt
> I think this is a regression caused by HDFS-127 -- DFSInputStream is supposed to only
go back to the NN max.block.acquires times, but in trunk it goes back twice as many - the
default is 3, but I am counting 7 calls to getBlockLocations before an exception is thrown.

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message