hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Colin Patrick McCabe (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-5957) Provide support for different mmap cache retention policies in ShortCircuitCache.
Date Fri, 21 Feb 2014 02:20:20 GMT

    [ https://issues.apache.org/jira/browse/HDFS-5957?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13907875#comment-13907875

Colin Patrick McCabe commented on HDFS-5957:

I talked to [~kkambatl] about this.  It seems that YARN is monitoring the process' {{RSS}}
(resident set size), which does seem to include the physical memory taken up by memory-mapped
files.  I think this is unfortunate.  The physical memory taken up by mmapped files is basically
part of the page cache.  If there is any memory pressure at all, it's easy to purge this memory
(the pages are "clean")  Charging an application for this memory is similar to charging it
for the page cache consumed by calls to read(2)-- it doesn't really make sense for this application.
 I think this is a problem within YARN, which has to be fixed inside YARN.

bq. It sounds like you really do need a deterministic way to trigger the munmap calls, i.e.
LRU caching or no caching at all described above.

The {{munmap}} calls are deterministic now.  You can control the number of unused mmaps that
we'll store by changing {{dfs.client.mmap.cache.size}}.

It's very important to keep in mind that {{dfs.client.mmap.cache.size}} controls the size
of the cache, *not* the total number of mmaps.  So if my application has 10 threads that each
use an mmap at a time, and the maximum cache size is 10, I may have 20 mmaps in existence
at any given time.  The maximum size of any mmap is going to be the size of a block, so you
should be able to use this to calculate how much RSS you will need.

bq. For small 200Gb data-sets (~1.4x tasks per container), ZCR does give a perf boost because
we get to use HADOOP-10047 instead of shuffling it between byte[] buffers for decompression.

As a workaround, have you considered reading into a direct {{ByteBuffer}} that you allocated
yourself?  {{DFSInputStream}} implements the {{ByteBufferReadable}} interface, which lets
you read into any {{ByteBuffer}}.  This would avoid the array copy that you're talking about.

I hope we can fix this within YARN soon, since otherwise the perf benefit of zero-copy reads
will be substantially reduced or eliminated (as well as people's ability to use ZCR in the
first place)

> Provide support for different mmap cache retention policies in ShortCircuitCache.
> ---------------------------------------------------------------------------------
>                 Key: HDFS-5957
>                 URL: https://issues.apache.org/jira/browse/HDFS-5957
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: hdfs-client
>    Affects Versions: 2.3.0
>            Reporter: Chris Nauroth
> Currently, the {{ShortCircuitCache}} retains {{mmap}} regions for reuse by multiple reads
of the same block or by multiple threads.  The eventual {{munmap}} executes on a background
thread after an expiration period.  Some client usage patterns would prefer strict bounds
on this cache and deterministic cleanup by calling {{munmap}}.  This issue proposes additional
support for different caching policies that better fit these usage patterns.

This message was sent by Atlassian JIRA

View raw message