hbase-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Enis Soztutar (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HBASE-10052) use HDFS advisory caching to avoid caching HFiles that are not going to be read again (because they are being compacted)
Date Wed, 04 Dec 2013 00:51:36 GMT

    [ https://issues.apache.org/jira/browse/HBASE-10052?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13838409#comment-13838409
] 

Enis Soztutar commented on HBASE-10052:
---------------------------------------

Sounds good. Any API we can use to direct hdfs to NOT cache our hfiles in the first place.
We are already caching the blocks, and it is better to have hdfs to not double cache, but
instead spend the memory elsewhere. Since only the HBase layer knows about block boundaries,
block access patterns and block types (index, data), HBase cache should always be better in
theory. 

> use HDFS advisory caching to avoid caching HFiles that are not going to be read again
(because they are being compacted)
> ------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HBASE-10052
>                 URL: https://issues.apache.org/jira/browse/HBASE-10052
>             Project: HBase
>          Issue Type: Improvement
>            Reporter: Colin Patrick McCabe
>             Fix For: 0.98.0
>
>
> HBase can benefit from doing dropbehind during compaction since compacted files are not
read again.  HDFS advisory caching, introduced in HDFS-4817, can help here.  The right API
here is {{DataInputStream#setDropBehind}}.



--
This message was sent by Atlassian JIRA
(v6.1#6144)

Mime
View raw message