hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Ashish Singhi (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-17290) Potential loss of data for replication of bulk loaded hfiles
Date Thu, 05 Jan 2017 17:01:58 GMT

    [ https://issues.apache.org/jira/browse/HBASE-17290?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15801873#comment-15801873

Ashish Singhi commented on HBASE-17290:

{quote} Can you point me to the code which handles the case where Path for bulk loaded hfile
is recorded but the commit (move of hfile) fails ?
In that scenario, the file wouldn't be found at time of replication. {quote}
In that scenario we will end up here, https://github.com/apache/hbase/blob/master/hbase-server/src/main/java/org/apache/hadoop/hbase/replication/regionserver/HFileReplicator.java#L380

> Potential loss of data for replication of bulk loaded hfiles
> ------------------------------------------------------------
>                 Key: HBASE-17290
>                 URL: https://issues.apache.org/jira/browse/HBASE-17290
>             Project: HBase
>          Issue Type: Bug
>    Affects Versions: 1.3.0
>            Reporter: Ted Yu
>            Assignee: Ashish Singhi
>             Fix For: 2.0.0, 1.4.0
>         Attachments: HBASE-17290.patch
> Currently the support for replication of bulk loaded hfiles relies on bulk load marker
written in the WAL.
> The move of bulk loaded hfile(s) (into region directory) may succeed but the write of
bulk load marker may fail.
> This means that although bulk loaded hfile is being served in source cluster, the replication
wouldn't happen.
> Normally operator is supposed to retry the bulk load. But relying on human retry is not
robust solution.

This message was sent by Atlassian JIRA

View raw message