hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "stack (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-7530) [replication] Work around HDFS-4380 else we get NPEs
Date Fri, 11 Jan 2013 00:08:13 GMT

    [ https://issues.apache.org/jira/browse/HBASE-7530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13550583#comment-13550583
] 

stack commented on HBASE-7530:
------------------------------

I don't get what this change does.  Previous we had an explicit sizing.   This does explicit
sizing too, right, by rolling at some multiple of current size?
                
> [replication] Work around HDFS-4380 else we get NPEs
> ----------------------------------------------------
>
>                 Key: HBASE-7530
>                 URL: https://issues.apache.org/jira/browse/HBASE-7530
>             Project: HBase
>          Issue Type: Bug
>    Affects Versions: 0.94.3
>            Reporter: Jean-Daniel Cryans
>            Assignee: Jean-Daniel Cryans
>             Fix For: 0.96.0, 0.94.5
>
>         Attachments: HBASE-7530.patch
>
>
> I've been spending a lot of time trying to figure the recent test failures related to
replication. One I seem to be constantly getting is this NPE:
> {noformat}
> 2013-01-09 10:08:56,912 ERROR [RegionServer:1;172.23.7.205,61604,1357754664830-EventThread.replicationSource,2]
regionserver.ReplicationSource$1(727): Unexpected exception in ReplicationSource, currentPath=hdfs://localhost:61589/user/jdcryans/hbase/.logs/172.23.7.205,61604,1357754664830/172.23.7.205%2C61604%2C1357754664830.1357754936216
> java.lang.NullPointerException
>         at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.updateBlockInfo(DFSClient.java:1885)
>         at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.openInfo(DFSClient.java:1858)
>         at org.apache.hadoop.hdfs.DFSClient$DFSInputStream.<init>(DFSClient.java:1834)
>         at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:578)
>         at org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSystem.java:154)
>         at org.apache.hadoop.fs.FilterFileSystem.open(FilterFileSystem.java:108)
>         at org.apache.hadoop.io.SequenceFile$Reader.openFile(SequenceFile.java:1495)
>         at org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogReader$WALReader.openFile(SequenceFileLogReader.java:62)
>         at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1482)
>         at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1475)
>         at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1470)
>         at org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogReader$WALReader.<init>(SequenceFileLogReader.java:55)
>         at org.apache.hadoop.hbase.regionserver.wal.SequenceFileLogReader.reset(SequenceFileLogReader.java:308)
>         at org.apache.hadoop.hbase.replication.regionserver.ReplicationHLogReaderManager.openReader(ReplicationHLogReaderManager.java:69)
>         at org.apache.hadoop.hbase.replication.regionserver.ReplicationSource.openReader(ReplicationSource.java:500)
>         at org.apache.hadoop.hbase.replication.regionserver.ReplicationSource.run(ReplicationSource.java:312)
> {noformat}
> Talking to [~tlipcon], he said it was likely fixed in Hadoop 2.0 via HDFS-3222 but for
Hadoop 1.0 he created HDFS-4380. This seems to happen while crossing block boundaries and
TestReplication uses a 20KB block size for the HLog. The intent was just to get HLogs to roll
more often, and this can also be achieved with *hbase.regionserver.logroll.multiplier* with
a value of 0.0003f.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message