hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "ramkrishna.s.vasudevan (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-18294) Flush is based on data size instead of heap size
Date Thu, 29 Jun 2017 15:13:00 GMT

    [ https://issues.apache.org/jira/browse/HBASE-18294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16068455#comment-16068455

ramkrishna.s.vasudevan commented on HBASE-18294:

In RegionServerAccounting we are trying to do all the flush calculation. Check 'isAboveHighWaterMark()'.
There we still consider the heapSize for the global heap pressure related flushes.
What is that you are observing now? Are you seeing some different behaviour now? Offheap memstores
need this change so that we are able to decide the region flush based on data size alone as
the whole data is offheap.

> Flush is based on data size instead of heap size
> ------------------------------------------------
>                 Key: HBASE-18294
>                 URL: https://issues.apache.org/jira/browse/HBASE-18294
>             Project: HBase
>          Issue Type: Bug
>            Reporter: Eshcar Hillel
>            Assignee: Eshcar Hillel
> A region is flushed if its memory component exceed a threshold (default size is 128MB).
> A flush policy decides whether to flush a store by comparing the size of the store to
another threshold (that can be configured with hbase.hregion.percolumnfamilyflush.size.lower.bound).
> Currently the implementation (in both cases) compares the data size (key-value only)
to the threshold where it should compare the heap size (which includes index size, and metadata).

This message was sent by Atlassian JIRA

View raw message